var/home/core/zuul-output/0000755000175000017500000000000015114032657014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114045221015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005317503715114045212017703 0ustar rootrootDec 03 13:09:13 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 13:09:13 crc restorecon[4646]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:09:13 crc restorecon[4646]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 13:09:14 crc kubenswrapper[4690]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 13:09:14 crc kubenswrapper[4690]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 13:09:14 crc kubenswrapper[4690]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 13:09:14 crc kubenswrapper[4690]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 13:09:14 crc kubenswrapper[4690]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 13:09:14 crc kubenswrapper[4690]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.133478 4690 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136565 4690 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136583 4690 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136588 4690 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136593 4690 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136597 4690 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136602 4690 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136607 4690 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136611 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136616 4690 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136619 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136623 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136626 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136630 4690 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136633 4690 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136637 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136653 4690 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136657 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136660 4690 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136664 4690 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136667 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136670 4690 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136674 4690 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136677 4690 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136680 4690 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136684 4690 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136687 4690 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136691 4690 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136694 4690 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136698 4690 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136701 4690 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136705 4690 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136709 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136714 4690 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136717 4690 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136721 4690 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136725 4690 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136729 4690 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136733 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136737 4690 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136740 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136744 4690 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136747 4690 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136750 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136754 4690 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136757 4690 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136760 4690 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136764 4690 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136767 4690 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136771 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136774 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136778 4690 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136787 4690 feature_gate.go:330] unrecognized feature gate: Example Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136791 4690 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136795 4690 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136798 4690 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136802 4690 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136806 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136809 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136813 4690 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136816 4690 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136821 4690 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136825 4690 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136829 4690 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136833 4690 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136837 4690 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136842 4690 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136846 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136850 4690 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136854 4690 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136858 4690 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.136875 4690 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.136952 4690 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.136960 4690 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.136984 4690 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.136990 4690 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.136996 4690 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137001 4690 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137007 4690 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137013 4690 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137017 4690 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137021 4690 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137026 4690 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137030 4690 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137034 4690 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137038 4690 flags.go:64] FLAG: --cgroup-root="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137042 4690 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137046 4690 flags.go:64] FLAG: --client-ca-file="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137053 4690 flags.go:64] FLAG: --cloud-config="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137058 4690 flags.go:64] FLAG: --cloud-provider="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137062 4690 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137067 4690 flags.go:64] FLAG: --cluster-domain="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137071 4690 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137075 4690 flags.go:64] FLAG: --config-dir="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137079 4690 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137084 4690 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137090 4690 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137095 4690 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137099 4690 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137103 4690 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137108 4690 flags.go:64] FLAG: --contention-profiling="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137112 4690 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137116 4690 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137120 4690 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137124 4690 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137129 4690 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137133 4690 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137137 4690 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137142 4690 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137145 4690 flags.go:64] FLAG: --enable-server="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137150 4690 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137155 4690 flags.go:64] FLAG: --event-burst="100" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137160 4690 flags.go:64] FLAG: --event-qps="50" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137165 4690 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137170 4690 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137175 4690 flags.go:64] FLAG: --eviction-hard="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137182 4690 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137186 4690 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137192 4690 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137197 4690 flags.go:64] FLAG: --eviction-soft="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137201 4690 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137206 4690 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137211 4690 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137215 4690 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137221 4690 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137226 4690 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137232 4690 flags.go:64] FLAG: --feature-gates="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137238 4690 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137244 4690 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137249 4690 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137254 4690 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137259 4690 flags.go:64] FLAG: --healthz-port="10248" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137264 4690 flags.go:64] FLAG: --help="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137269 4690 flags.go:64] FLAG: --hostname-override="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137274 4690 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137279 4690 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137284 4690 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137289 4690 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137294 4690 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137300 4690 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137305 4690 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137309 4690 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137313 4690 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137317 4690 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137322 4690 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137326 4690 flags.go:64] FLAG: --kube-reserved="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137330 4690 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137334 4690 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137338 4690 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137342 4690 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137346 4690 flags.go:64] FLAG: --lock-file="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137350 4690 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137354 4690 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137358 4690 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137364 4690 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137369 4690 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137373 4690 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137377 4690 flags.go:64] FLAG: --logging-format="text" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137381 4690 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137386 4690 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137391 4690 flags.go:64] FLAG: --manifest-url="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137396 4690 flags.go:64] FLAG: --manifest-url-header="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137402 4690 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137405 4690 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137410 4690 flags.go:64] FLAG: --max-pods="110" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137415 4690 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137419 4690 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137423 4690 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137427 4690 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137431 4690 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137436 4690 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137440 4690 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137451 4690 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137455 4690 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137459 4690 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137463 4690 flags.go:64] FLAG: --pod-cidr="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137467 4690 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137474 4690 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137478 4690 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137482 4690 flags.go:64] FLAG: --pods-per-core="0" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137486 4690 flags.go:64] FLAG: --port="10250" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137491 4690 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137495 4690 flags.go:64] FLAG: --provider-id="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137499 4690 flags.go:64] FLAG: --qos-reserved="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137502 4690 flags.go:64] FLAG: --read-only-port="10255" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137506 4690 flags.go:64] FLAG: --register-node="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137511 4690 flags.go:64] FLAG: --register-schedulable="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137515 4690 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137522 4690 flags.go:64] FLAG: --registry-burst="10" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137526 4690 flags.go:64] FLAG: --registry-qps="5" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137530 4690 flags.go:64] FLAG: --reserved-cpus="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137534 4690 flags.go:64] FLAG: --reserved-memory="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137539 4690 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137543 4690 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137547 4690 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137551 4690 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137557 4690 flags.go:64] FLAG: --runonce="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137561 4690 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137565 4690 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137570 4690 flags.go:64] FLAG: --seccomp-default="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137574 4690 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137578 4690 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137584 4690 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137588 4690 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137593 4690 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137597 4690 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137601 4690 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137606 4690 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137610 4690 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137614 4690 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137618 4690 flags.go:64] FLAG: --system-cgroups="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137622 4690 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137628 4690 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137632 4690 flags.go:64] FLAG: --tls-cert-file="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137636 4690 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137648 4690 flags.go:64] FLAG: --tls-min-version="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137652 4690 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137656 4690 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137660 4690 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137664 4690 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137668 4690 flags.go:64] FLAG: --v="2" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137674 4690 flags.go:64] FLAG: --version="false" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137680 4690 flags.go:64] FLAG: --vmodule="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137686 4690 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.137690 4690 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137843 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137850 4690 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137879 4690 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137886 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137890 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137895 4690 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137899 4690 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137904 4690 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137908 4690 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137911 4690 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137918 4690 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137922 4690 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137925 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137929 4690 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137933 4690 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137936 4690 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137940 4690 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137943 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137947 4690 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137951 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137954 4690 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137957 4690 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137961 4690 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137965 4690 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137968 4690 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137972 4690 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137975 4690 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137979 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137982 4690 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137986 4690 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137990 4690 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137993 4690 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.137997 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138000 4690 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138004 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138007 4690 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138011 4690 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138015 4690 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138019 4690 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138022 4690 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138026 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138029 4690 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138036 4690 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138041 4690 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138046 4690 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138050 4690 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138053 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138058 4690 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138062 4690 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138067 4690 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138071 4690 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138075 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138079 4690 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138083 4690 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138087 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138092 4690 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138095 4690 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138099 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138103 4690 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138106 4690 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138110 4690 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138113 4690 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138117 4690 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138121 4690 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138124 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138128 4690 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138132 4690 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138136 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138140 4690 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138144 4690 feature_gate.go:330] unrecognized feature gate: Example Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.138148 4690 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.138300 4690 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.147487 4690 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.147856 4690 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148072 4690 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148095 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148105 4690 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148114 4690 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148124 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148132 4690 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148140 4690 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148148 4690 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148156 4690 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148164 4690 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148171 4690 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148179 4690 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148186 4690 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148194 4690 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148202 4690 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148210 4690 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148217 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148226 4690 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148233 4690 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148241 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148249 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148257 4690 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148265 4690 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148272 4690 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148285 4690 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148297 4690 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148308 4690 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148317 4690 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148336 4690 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148344 4690 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148352 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148360 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148370 4690 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148380 4690 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148388 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148396 4690 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148404 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148412 4690 feature_gate.go:330] unrecognized feature gate: Example Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148419 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148427 4690 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148435 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148442 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148452 4690 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148461 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148469 4690 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148477 4690 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148485 4690 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148492 4690 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148500 4690 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148508 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148516 4690 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148524 4690 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148533 4690 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148541 4690 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148549 4690 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148557 4690 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148564 4690 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148572 4690 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148579 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148587 4690 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148596 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148604 4690 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148612 4690 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148624 4690 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148634 4690 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148643 4690 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148653 4690 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148661 4690 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148669 4690 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148677 4690 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148685 4690 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.148699 4690 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148957 4690 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148972 4690 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148981 4690 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.148989 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149000 4690 feature_gate.go:330] unrecognized feature gate: Example Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149008 4690 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149017 4690 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149026 4690 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149034 4690 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149043 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149051 4690 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149059 4690 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149067 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149075 4690 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149084 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149091 4690 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149099 4690 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149107 4690 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149114 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149123 4690 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149131 4690 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149139 4690 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149146 4690 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149157 4690 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149168 4690 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149177 4690 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149186 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149194 4690 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149202 4690 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149210 4690 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149219 4690 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149227 4690 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149235 4690 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149244 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149252 4690 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149260 4690 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149269 4690 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149280 4690 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149289 4690 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149299 4690 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149309 4690 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149319 4690 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149328 4690 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149336 4690 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149344 4690 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149353 4690 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149360 4690 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149368 4690 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149376 4690 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149384 4690 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149392 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149400 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149408 4690 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149416 4690 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149424 4690 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149432 4690 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149442 4690 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149451 4690 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149459 4690 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149467 4690 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149474 4690 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149483 4690 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149490 4690 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149498 4690 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149506 4690 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149514 4690 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149522 4690 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149529 4690 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149537 4690 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149544 4690 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.149552 4690 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.149565 4690 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.150180 4690 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.154359 4690 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.154500 4690 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.155737 4690 server.go:997] "Starting client certificate rotation" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.155788 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.156056 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-08 22:22:01.012855784 +0000 UTC Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.156202 4690 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 873h12m46.856658799s for next certificate rotation Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.162309 4690 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.164194 4690 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.174674 4690 log.go:25] "Validated CRI v1 runtime API" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.199300 4690 log.go:25] "Validated CRI v1 image API" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.201401 4690 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.204504 4690 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-13-04-53-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.204557 4690 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.232372 4690 manager.go:217] Machine: {Timestamp:2025-12-03 13:09:14.231080998 +0000 UTC m=+0.212001451 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c6db3655-bd43-479c-8c13-6f91f454fdd0 BootID:95a051aa-8ddb-4533-bbde-548417042fa3 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:50:0e:2b Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:50:0e:2b Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:76:a7:86 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e2:05:21 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:4f:3c:dd Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:44:6e:86 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8a:69:59:7c:17:e8 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ba:1b:5b:42:47:25 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.232593 4690 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.232904 4690 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.233744 4690 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.233977 4690 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.234023 4690 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.234240 4690 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.234250 4690 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.234451 4690 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.234488 4690 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.234701 4690 state_mem.go:36] "Initialized new in-memory state store" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.234800 4690 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.235757 4690 kubelet.go:418] "Attempting to sync node with API server" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.235778 4690 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.235807 4690 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.235821 4690 kubelet.go:324] "Adding apiserver pod source" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.235832 4690 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.238183 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.238184 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.238502 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.238923 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.238986 4690 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.239404 4690 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240159 4690 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240733 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240760 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240767 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240774 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240788 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240795 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240804 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240820 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240831 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240840 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240891 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.240907 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.241260 4690 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.241769 4690 server.go:1280] "Started kubelet" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.242304 4690 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.242296 4690 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.242807 4690 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.243189 4690 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 13:09:14 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.244171 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.244193 4690 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.245537 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 00:53:33.660620383 +0000 UTC Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.247993 4690 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.248022 4690 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.248077 4690 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.248149 4690 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.246809 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="200ms" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.248480 4690 server.go:460] "Adding debug handlers to kubelet server" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.248560 4690 factory.go:55] Registering systemd factory Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.248608 4690 factory.go:221] Registration of the systemd container factory successfully Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.249406 4690 factory.go:153] Registering CRI-O factory Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.249443 4690 factory.go:221] Registration of the crio container factory successfully Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.249428 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.249482 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.255070 4690 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.255129 4690 factory.go:103] Registering Raw factory Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.255148 4690 manager.go:1196] Started watching for new ooms in manager Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.256664 4690 manager.go:319] Starting recovery of all containers Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.256060 4690 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187db68efcc8a1d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 13:09:14.241720792 +0000 UTC m=+0.222641245,LastTimestamp:2025-12-03 13:09:14.241720792 +0000 UTC m=+0.222641245,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.267985 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268037 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268052 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268062 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268071 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268088 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268099 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268108 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268122 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268131 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268141 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268150 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268161 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268173 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268187 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268198 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268206 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268215 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268226 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268256 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268265 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268273 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268282 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268291 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268301 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268310 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268322 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268350 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268359 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268387 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268398 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268407 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268417 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268426 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268435 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268444 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268456 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268466 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268475 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268485 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268495 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268505 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268514 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268524 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268534 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268544 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268555 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268565 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268575 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268586 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.268598 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269137 4690 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269161 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269177 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269191 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269201 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269212 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269224 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269233 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269246 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269257 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269268 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269278 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269288 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269298 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269308 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269318 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269329 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269340 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269351 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269362 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269375 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269386 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269396 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269407 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269418 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269429 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269439 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269450 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269461 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269473 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269483 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269494 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269504 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269516 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269527 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269540 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269550 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269559 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269569 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269578 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269588 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269599 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269609 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269619 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269633 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269656 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269671 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269682 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269695 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269715 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269728 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269747 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269760 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269772 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269826 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269841 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269852 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269862 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269886 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269898 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269909 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269920 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269932 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269942 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269952 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269963 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269972 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269982 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.269991 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270000 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270011 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270023 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270035 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270047 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270059 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270070 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270080 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270094 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270109 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270121 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270135 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270149 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270158 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270173 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270182 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270193 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270202 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270212 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270221 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270231 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270239 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270250 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270259 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270271 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270280 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270288 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270297 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270308 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270317 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270328 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270338 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270347 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270357 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270367 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270378 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270388 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270397 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270406 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270415 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270424 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270432 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270441 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270449 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270460 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270468 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270479 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270489 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270498 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270506 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270515 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270524 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270534 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270542 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270551 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270560 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270569 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270577 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270588 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270596 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270605 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270615 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270626 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270635 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270645 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270653 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270662 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270671 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270681 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270690 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270700 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270710 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270720 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270731 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270739 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270747 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270755 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270764 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270775 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270783 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270793 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270802 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270811 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270820 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270828 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270836 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270849 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270857 4690 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270884 4690 reconstruct.go:97] "Volume reconstruction finished" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.270892 4690 reconciler.go:26] "Reconciler: start to sync state" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.295382 4690 manager.go:324] Recovery completed Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.310110 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.310549 4690 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.313099 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.313143 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.313154 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.313385 4690 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.313448 4690 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.313489 4690 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.313562 4690 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.314334 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.314407 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.315007 4690 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.315035 4690 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.315061 4690 state_mem.go:36] "Initialized new in-memory state store" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.325168 4690 policy_none.go:49] "None policy: Start" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.326192 4690 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.326225 4690 state_mem.go:35] "Initializing new in-memory state store" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.348342 4690 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.385341 4690 manager.go:334] "Starting Device Plugin manager" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.385427 4690 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.385441 4690 server.go:79] "Starting device plugin registration server" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.386022 4690 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.386044 4690 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.389570 4690 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.389668 4690 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.389681 4690 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.398650 4690 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.414295 4690 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.414604 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.415905 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.416088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.416182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.416435 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.416711 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.416801 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.417489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.417624 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.417710 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.417918 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.418184 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.418274 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.418201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.418353 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.418370 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419239 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419329 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419504 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419656 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419711 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419751 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.419826 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.420801 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.420936 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.420839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.421095 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.421117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.421055 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.421395 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.421512 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.421548 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.422499 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.422531 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.422543 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.422709 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.422740 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.422989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.423031 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.423042 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.423652 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.423817 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.424067 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.449510 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="400ms" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474043 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474130 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474187 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474223 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474264 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474300 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474333 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474369 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474404 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474436 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474468 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474502 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474542 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474587 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.474632 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.486169 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.487649 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.487733 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.487764 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.487835 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.488594 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576151 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576228 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576264 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576299 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576327 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576354 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576376 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576427 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576477 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576454 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576561 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576597 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576628 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576631 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576651 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576628 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576678 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576676 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576721 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576722 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576770 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576599 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576771 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576811 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576784 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576858 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576896 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576742 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576953 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.576830 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.689170 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.690983 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.691057 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.691085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.691133 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.691673 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.745388 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.774346 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.779463 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-0ef555ab5c6be88950ae77d4231e01db1f7877f1f51630091ecfc15f1b021bbf WatchSource:0}: Error finding container 0ef555ab5c6be88950ae77d4231e01db1f7877f1f51630091ecfc15f1b021bbf: Status 404 returned error can't find the container with id 0ef555ab5c6be88950ae77d4231e01db1f7877f1f51630091ecfc15f1b021bbf Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.794068 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.801668 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.804162 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-380497c92803e9ccbcc61e39b248834d6bcaf8c0cba8f713a55439c41b97cd2e WatchSource:0}: Error finding container 380497c92803e9ccbcc61e39b248834d6bcaf8c0cba8f713a55439c41b97cd2e: Status 404 returned error can't find the container with id 380497c92803e9ccbcc61e39b248834d6bcaf8c0cba8f713a55439c41b97cd2e Dec 03 13:09:14 crc kubenswrapper[4690]: I1203 13:09:14.807208 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:09:14 crc kubenswrapper[4690]: W1203 13:09:14.826288 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-517c05507b999022287722e3b086165095f030780ff5ca6800fa4bf55d812255 WatchSource:0}: Error finding container 517c05507b999022287722e3b086165095f030780ff5ca6800fa4bf55d812255: Status 404 returned error can't find the container with id 517c05507b999022287722e3b086165095f030780ff5ca6800fa4bf55d812255 Dec 03 13:09:14 crc kubenswrapper[4690]: E1203 13:09:14.851330 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="800ms" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.091955 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.093797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.093857 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.093888 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.093927 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:09:15 crc kubenswrapper[4690]: E1203 13:09:15.094509 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.244515 4690 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:15 crc kubenswrapper[4690]: W1203 13:09:15.245932 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:15 crc kubenswrapper[4690]: E1203 13:09:15.245997 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.246037 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 16:08:44.238439622 +0000 UTC Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.246074 4690 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 146h59m28.992368123s for next certificate rotation Dec 03 13:09:15 crc kubenswrapper[4690]: W1203 13:09:15.262277 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:15 crc kubenswrapper[4690]: E1203 13:09:15.262365 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.323538 4690 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f" exitCode=0 Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.323640 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.323753 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bd7a1d75021f3f0f07146751111b4cc954deb2ae509a631faa391489bd48a94e"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.323899 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.325461 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.325492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.325504 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.326987 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.327104 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"517c05507b999022287722e3b086165095f030780ff5ca6800fa4bf55d812255"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.329661 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f" exitCode=0 Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.329721 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.329757 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"380497c92803e9ccbcc61e39b248834d6bcaf8c0cba8f713a55439c41b97cd2e"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.329857 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.331073 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.331133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.331158 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.332089 4690 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850" exitCode=0 Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.332128 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.332177 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0ef555ab5c6be88950ae77d4231e01db1f7877f1f51630091ecfc15f1b021bbf"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.332333 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.333166 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.333309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.333335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.333348 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.333904 4690 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2" exitCode=0 Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.333951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.333956 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.334035 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ae67656334db569dd426cd47be8da8f3a4eb0694de290bbb26dbf86ab490c670"} Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.333995 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.334094 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.334148 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.335134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.335180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.335203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:15 crc kubenswrapper[4690]: E1203 13:09:15.652994 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="1.6s" Dec 03 13:09:15 crc kubenswrapper[4690]: W1203 13:09:15.687994 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:15 crc kubenswrapper[4690]: E1203 13:09:15.688108 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:09:15 crc kubenswrapper[4690]: W1203 13:09:15.734441 4690 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 03 13:09:15 crc kubenswrapper[4690]: E1203 13:09:15.734775 4690 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.895645 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.897328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.897384 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.897395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:15 crc kubenswrapper[4690]: I1203 13:09:15.897428 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:09:15 crc kubenswrapper[4690]: E1203 13:09:15.897844 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.338579 4690 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada" exitCode=0 Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.338641 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.338752 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.339558 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.339580 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.339599 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.348361 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"de9a2357cbe74fa94cec46032822a70d9124c382e568d434ec7dfc64167aa8aa"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.348489 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.349428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.349453 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.349461 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.354982 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.355025 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.355038 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.355135 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.356215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.356238 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.356249 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.362647 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.362688 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.362699 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.362795 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.368088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.368134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.368157 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.371144 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.371170 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.371182 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539"} Dec 03 13:09:16 crc kubenswrapper[4690]: I1203 13:09:16.371191 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984"} Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.379248 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff"} Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.382299 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.384244 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.384290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.384317 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.389118 4690 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587" exitCode=0 Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.389199 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587"} Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.389337 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.389363 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.390644 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.390706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.390723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.391606 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.391661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.391672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.498372 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.499804 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.499848 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.499860 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.499909 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.714502 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.714755 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.716450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.716537 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:17 crc kubenswrapper[4690]: I1203 13:09:17.716553 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.395709 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9"} Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.395762 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03"} Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.395776 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a"} Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.395786 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4"} Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.395845 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.395928 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.396985 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.397008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:18 crc kubenswrapper[4690]: I1203 13:09:18.397017 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.242435 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.242568 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.246380 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.246432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.246443 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.251055 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.401415 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6"} Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.401440 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.401440 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.402395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.402432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.402442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.402642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.402679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.402691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.744698 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.786219 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.786380 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.786432 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.788070 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.788109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:19 crc kubenswrapper[4690]: I1203 13:09:19.788124 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.403421 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.403485 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.404647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.404691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.404706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.404973 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.405025 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.405040 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:20 crc kubenswrapper[4690]: I1203 13:09:20.652596 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.313156 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.313331 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.313372 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.314728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.314767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.314779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.323130 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.406167 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.406295 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.407330 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.407361 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.407372 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.407678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.407719 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.407738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:21 crc kubenswrapper[4690]: I1203 13:09:21.761054 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.409164 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.410262 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.410300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.410311 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.820858 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.821168 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.823495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.823545 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:22 crc kubenswrapper[4690]: I1203 13:09:22.823558 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:24 crc kubenswrapper[4690]: I1203 13:09:24.154668 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:24 crc kubenswrapper[4690]: I1203 13:09:24.154985 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:24 crc kubenswrapper[4690]: I1203 13:09:24.156594 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:24 crc kubenswrapper[4690]: I1203 13:09:24.156629 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:24 crc kubenswrapper[4690]: I1203 13:09:24.156641 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:24 crc kubenswrapper[4690]: E1203 13:09:24.399578 4690 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 13:09:26 crc kubenswrapper[4690]: I1203 13:09:26.245477 4690 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 13:09:26 crc kubenswrapper[4690]: I1203 13:09:26.995485 4690 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 13:09:26 crc kubenswrapper[4690]: I1203 13:09:26.995581 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 13:09:27 crc kubenswrapper[4690]: I1203 13:09:27.002316 4690 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 13:09:27 crc kubenswrapper[4690]: I1203 13:09:27.002426 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 13:09:27 crc kubenswrapper[4690]: I1203 13:09:27.155256 4690 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 13:09:27 crc kubenswrapper[4690]: I1203 13:09:27.155338 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 13:09:29 crc kubenswrapper[4690]: I1203 13:09:29.750707 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:29 crc kubenswrapper[4690]: I1203 13:09:29.752312 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:29 crc kubenswrapper[4690]: I1203 13:09:29.754202 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:29 crc kubenswrapper[4690]: I1203 13:09:29.754423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:29 crc kubenswrapper[4690]: I1203 13:09:29.754608 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.319636 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.319972 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.322421 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.322477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.322540 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.325469 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.432625 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.433837 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.433917 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.433931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.790305 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.790966 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.792691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.792763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.792797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.806936 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 13:09:31 crc kubenswrapper[4690]: E1203 13:09:31.987659 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.992482 4690 trace.go:236] Trace[1366398713]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 13:09:17.702) (total time: 14289ms): Dec 03 13:09:31 crc kubenswrapper[4690]: Trace[1366398713]: ---"Objects listed" error: 14289ms (13:09:31.992) Dec 03 13:09:31 crc kubenswrapper[4690]: Trace[1366398713]: [14.289906408s] [14.289906408s] END Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.992529 4690 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.993304 4690 trace.go:236] Trace[1412426570]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 13:09:18.406) (total time: 13586ms): Dec 03 13:09:31 crc kubenswrapper[4690]: Trace[1412426570]: ---"Objects listed" error: 13586ms (13:09:31.993) Dec 03 13:09:31 crc kubenswrapper[4690]: Trace[1412426570]: [13.586808681s] [13.586808681s] END Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.993335 4690 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.993962 4690 trace.go:236] Trace[46628265]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 13:09:17.368) (total time: 14625ms): Dec 03 13:09:31 crc kubenswrapper[4690]: Trace[46628265]: ---"Objects listed" error: 14625ms (13:09:31.993) Dec 03 13:09:31 crc kubenswrapper[4690]: Trace[46628265]: [14.625496959s] [14.625496959s] END Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.994186 4690 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.995734 4690 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 13:09:31 crc kubenswrapper[4690]: E1203 13:09:31.996172 4690 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.996583 4690 trace.go:236] Trace[1521431785]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 13:09:18.676) (total time: 13319ms): Dec 03 13:09:31 crc kubenswrapper[4690]: Trace[1521431785]: ---"Objects listed" error: 13319ms (13:09:31.996) Dec 03 13:09:31 crc kubenswrapper[4690]: Trace[1521431785]: [13.319739829s] [13.319739829s] END Dec 03 13:09:31 crc kubenswrapper[4690]: I1203 13:09:31.996612 4690 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.245661 4690 apiserver.go:52] "Watching apiserver" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.250124 4690 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.250423 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.251055 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.251222 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.251293 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.251826 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.251984 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.252101 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.251843 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.251826 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.252437 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.253953 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.254147 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.254590 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.254707 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.255008 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.255201 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.255271 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.255423 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.255575 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.264575 4690 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33184->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.264640 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33184->192.168.126.11:17697: read: connection reset by peer" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.265022 4690 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.265044 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.293635 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.319928 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.333505 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.343393 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.348780 4690 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.395789 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.397965 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398000 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398018 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398035 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398058 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398377 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398384 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398430 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398447 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398464 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398458 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398510 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398531 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398550 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398567 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398584 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398598 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398604 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398635 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398655 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398669 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398686 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398708 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398731 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398755 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398783 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398807 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398827 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398888 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398915 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398952 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398977 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.398995 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399011 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399027 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399044 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399061 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399119 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399144 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399167 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399478 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399484 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399755 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399886 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.399902 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400047 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400074 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400274 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400301 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400340 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400260 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400489 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400532 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400691 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400779 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400852 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400930 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.400988 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401022 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401113 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401183 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401229 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401265 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401264 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401328 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401408 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401426 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401465 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401548 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401603 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401639 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401663 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401689 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401714 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401738 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401762 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401784 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401806 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401827 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401854 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401900 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401932 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401954 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401976 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.401998 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402022 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402049 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402072 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402095 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402119 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402141 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402162 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402187 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402216 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402239 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402261 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402286 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402309 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402331 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402352 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402375 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402396 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402422 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402443 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402465 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404994 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405127 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405228 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.406752 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407800 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407842 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407880 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407907 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407974 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408009 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408028 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408053 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408078 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408099 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408118 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408137 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408157 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408176 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408192 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408210 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408227 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408243 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408261 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408280 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408298 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408328 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408345 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408361 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408380 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408397 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408413 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408458 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408476 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408496 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408514 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408532 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408549 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408565 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408625 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408648 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408664 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408682 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408699 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408717 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408734 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408749 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408766 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408784 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408849 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408909 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408927 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408946 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408965 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408983 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409000 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409018 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409036 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409062 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409080 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409096 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409112 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409129 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409148 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409167 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409183 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409199 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409215 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409231 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409247 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409263 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409280 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409299 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409319 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409338 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409358 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409377 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409398 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409414 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409431 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409450 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409466 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409483 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409499 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409516 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409532 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409549 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409564 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409579 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409596 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409614 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409630 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409648 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409664 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410292 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410328 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410356 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410416 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410438 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410466 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410491 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410512 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.410536 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402681 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.402996 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.403366 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.403670 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.403695 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.403704 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.403845 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404179 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404198 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404214 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404431 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404747 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404764 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404825 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.404896 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405029 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405067 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405093 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405290 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405324 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405671 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405708 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405721 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.405793 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.406058 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.406276 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.406287 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.406567 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.406693 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.406774 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407189 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407200 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407353 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407359 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.407567 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408375 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408396 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408293 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408444 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408787 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408889 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.408909 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409026 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409275 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409324 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409377 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.409625 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.411401 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.411828 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.412776 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.413387 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.413856 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.414228 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.414463 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.414908 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.416290 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.417808 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.418061 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.418508 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.418820 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.418915 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.418917 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419159 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419359 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419360 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419550 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419706 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419764 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419800 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419828 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419851 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419887 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419912 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419935 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.419956 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420001 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420027 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420003 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420054 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420078 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420085 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420106 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420129 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420148 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420170 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420191 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420213 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.423160 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420102 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.427351 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.427406 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.427411 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420476 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.427438 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420485 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420666 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420713 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420766 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.421019 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.420357 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.427560 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.421533 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.421357 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.421593 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.421830 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.421940 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.422195 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.422018 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.422530 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:09:32.922497262 +0000 UTC m=+18.903417695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.422779 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.422821 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.423067 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.422668 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.423203 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.423255 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.423490 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.423670 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.424694 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.425203 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.425205 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.425337 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.425757 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.425995 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.426017 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.426189 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.426616 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.426700 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.426835 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.427053 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.427128 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.427169 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.421194 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428051 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428112 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428175 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428211 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428242 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428263 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428286 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428306 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428329 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428350 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428374 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428407 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428402 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428513 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428536 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428608 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428654 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428657 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428792 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428814 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428833 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428902 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428944 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428967 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428981 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.428995 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429009 4690 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429023 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429037 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429053 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429066 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429085 4690 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429098 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429114 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429127 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429141 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429156 4690 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429169 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429186 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429198 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429210 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429223 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429236 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429249 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429262 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429275 4690 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429285 4690 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429296 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429316 4690 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429325 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429333 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429342 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429342 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429430 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429353 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429496 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429506 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429452 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429759 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429788 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429891 4690 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430104 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430224 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430250 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430296 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430393 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430538 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.429513 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430604 4690 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430624 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430638 4690 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430653 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430665 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430677 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430688 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430700 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430712 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430722 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430732 4690 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430742 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430754 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430806 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430819 4690 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430829 4690 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430840 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430850 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430861 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430887 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430898 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430909 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430918 4690 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430929 4690 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430939 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430950 4690 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430961 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430971 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430982 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430992 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431003 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431012 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431024 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431034 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431043 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431053 4690 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431064 4690 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431074 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431083 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431093 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431106 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431117 4690 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431127 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431137 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431165 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431175 4690 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431186 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431198 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431207 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431218 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431230 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.430843 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431146 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431277 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431394 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.431475 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.431535 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:32.931515496 +0000 UTC m=+18.912435929 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431578 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431698 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.431786 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431808 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.431978 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.432035 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:32.932027339 +0000 UTC m=+18.912947772 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432057 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432069 4690 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432081 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432095 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432112 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432140 4690 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432161 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432174 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432185 4690 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432196 4690 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.432207 4690 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.434607 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.434771 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.435107 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.435107 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.436850 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.439521 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.440027 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.440521 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.442072 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.445650 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.445956 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.446568 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.449833 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.449880 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.449896 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.449965 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:32.949941973 +0000 UTC m=+18.930862596 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.450203 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.450255 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.450267 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.450321 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:32.950291932 +0000 UTC m=+18.931212365 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.450208 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff" exitCode=255 Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.450640 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff"} Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.450658 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.452606 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.452718 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.453199 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.454509 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.456001 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.456265 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.457688 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.459908 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.461401 4690 scope.go:117] "RemoveContainer" containerID="85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.461571 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.462768 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.466662 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.468268 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.468659 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.468752 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.469162 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.471330 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.471439 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.471478 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.471571 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.471854 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.471966 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.472133 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.472142 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.472179 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.472797 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.472946 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.472998 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.473058 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.475227 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.480490 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.483787 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.483838 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.489278 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.496453 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.501507 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.507233 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.517211 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.529182 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.533627 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.533743 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534053 4690 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534087 4690 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534114 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534138 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534156 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534169 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534197 4690 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534213 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534228 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534249 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534264 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534282 4690 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534303 4690 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534331 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534348 4690 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534365 4690 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534384 4690 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534407 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534425 4690 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534442 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534466 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534484 4690 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534500 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534517 4690 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534538 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534553 4690 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534567 4690 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534580 4690 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534602 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534616 4690 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534631 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534648 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534673 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534688 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534702 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534723 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534738 4690 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534753 4690 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534768 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534818 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534832 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534846 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534860 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534899 4690 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534911 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534924 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534937 4690 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534957 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534975 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.534991 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535013 4690 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535027 4690 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535042 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535055 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535077 4690 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535092 4690 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535108 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535122 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535143 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535162 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535177 4690 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535195 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535210 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535274 4690 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535293 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535316 4690 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535331 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535345 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535358 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535381 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535400 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535419 4690 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535432 4690 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535453 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535467 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535484 4690 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535504 4690 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535518 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535532 4690 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535545 4690 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535569 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535584 4690 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535599 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535612 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535633 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535647 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535660 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535677 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535689 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535702 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535714 4690 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.535735 4690 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.536007 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.536156 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.541948 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.551728 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.564771 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.571521 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.579999 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.585482 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.594921 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:09:32 crc kubenswrapper[4690]: W1203 13:09:32.596167 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-36a4f8621133dbaf96ce6d2cf3a8f49582b5b7dad76bb0cc54f76bcccd4e06d9 WatchSource:0}: Error finding container 36a4f8621133dbaf96ce6d2cf3a8f49582b5b7dad76bb0cc54f76bcccd4e06d9: Status 404 returned error can't find the container with id 36a4f8621133dbaf96ce6d2cf3a8f49582b5b7dad76bb0cc54f76bcccd4e06d9 Dec 03 13:09:32 crc kubenswrapper[4690]: W1203 13:09:32.625900 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-ccca8c83937e68c8065df3770302584a72582136d8479ec8b1ff4c0958935899 WatchSource:0}: Error finding container ccca8c83937e68c8065df3770302584a72582136d8479ec8b1ff4c0958935899: Status 404 returned error can't find the container with id ccca8c83937e68c8065df3770302584a72582136d8479ec8b1ff4c0958935899 Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.939593 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.939668 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:32 crc kubenswrapper[4690]: I1203 13:09:32.939711 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.939786 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.939839 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:33.939826029 +0000 UTC m=+19.920746462 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.940258 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.940267 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:09:33.94024287 +0000 UTC m=+19.921163303 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:09:32 crc kubenswrapper[4690]: E1203 13:09:32.940346 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:33.940332542 +0000 UTC m=+19.921253025 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.040848 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.040965 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.041066 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.041113 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.041129 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.041136 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.041158 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.041178 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.041202 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:34.041181144 +0000 UTC m=+20.022101637 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.041247 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:34.041227165 +0000 UTC m=+20.022147768 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.314211 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.314364 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.455834 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.458575 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c"} Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.458908 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.459909 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ccca8c83937e68c8065df3770302584a72582136d8479ec8b1ff4c0958935899"} Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.462002 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848"} Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.462043 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5"} Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.462054 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"36a4f8621133dbaf96ce6d2cf3a8f49582b5b7dad76bb0cc54f76bcccd4e06d9"} Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.463382 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5"} Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.463403 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5d2fcfd4efc10315972d7709c4df8bf55fbb1d9e22933110c868fa2f06c3fc4f"} Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.498857 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.523280 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.554060 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.579797 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.597774 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.617665 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.659080 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.679656 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.693723 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.736662 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.754030 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.773211 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.787667 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.804190 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.820146 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.842789 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.949618 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.949719 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:33 crc kubenswrapper[4690]: I1203 13:09:33.949752 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.949983 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.950063 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:35.950041746 +0000 UTC m=+21.930962179 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.950498 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:09:35.950488877 +0000 UTC m=+21.931409310 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.950544 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:33 crc kubenswrapper[4690]: E1203 13:09:33.950568 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:35.950562359 +0000 UTC m=+21.931482792 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.050391 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.050431 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.050544 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.050558 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.050569 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.050613 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:36.050600061 +0000 UTC m=+22.031520484 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.050913 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.050926 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.050933 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.050954 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:36.05094784 +0000 UTC m=+22.031868273 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.142111 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-qh4c5"] Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.142534 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-p5fc8"] Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.142719 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qh4c5" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.142823 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.142742 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-dthlv"] Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.143646 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-tf2cb"] Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.143798 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.144198 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.145395 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vmddn"] Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.146286 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.147268 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.147271 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.147278 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.149525 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.149767 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.149835 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.149942 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.149958 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.150529 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.150680 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.151009 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.151201 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.152155 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158166 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158207 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158255 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158278 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158291 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158291 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158416 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158431 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.158458 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.162264 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.173172 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.182240 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.198733 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.209858 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.211724 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.227084 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.243657 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252270 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-k8s-cni-cncf-io\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252324 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-kubelet\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252361 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-systemd-units\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252393 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-netd\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252474 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-system-cni-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252532 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-cni-bin\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252555 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-proxy-tls\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252577 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/545e942f-5352-4a89-8e75-17762d512364-cni-binary-copy\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252606 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-node-log\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252636 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-os-release\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252660 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-etc-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252678 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-netns\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252698 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-multus-certs\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252749 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-os-release\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252769 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-systemd\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252787 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-env-overrides\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252848 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-hostroot\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252900 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-bin\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252935 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7hgf\" (UniqueName: \"kubernetes.io/projected/c8490fb8-480e-4d81-82dd-4411b1f29278-kube-api-access-r7hgf\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.252984 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bcf22e98-c7f6-4eb3-9c13-60f1953a9271-hosts-file\") pod \"node-resolver-qh4c5\" (UID: \"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\") " pod="openshift-dns/node-resolver-qh4c5" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253008 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-rootfs\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253026 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253119 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwc65\" (UniqueName: \"kubernetes.io/projected/bf580211-e03a-444e-98b8-c12f99091b80-kube-api-access-nwc65\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253156 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253184 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-script-lib\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253215 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-log-socket\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253275 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-cnibin\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253307 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-kubelet\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253330 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-etc-kubernetes\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253361 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxp5r\" (UniqueName: \"kubernetes.io/projected/545e942f-5352-4a89-8e75-17762d512364-kube-api-access-qxp5r\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253386 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-ovn-kubernetes\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253439 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-socket-dir-parent\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253467 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-conf-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253497 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-cnibin\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253547 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpngs\" (UniqueName: \"kubernetes.io/projected/bcf22e98-c7f6-4eb3-9c13-60f1953a9271-kube-api-access-tpngs\") pod \"node-resolver-qh4c5\" (UID: \"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\") " pod="openshift-dns/node-resolver-qh4c5" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253577 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253610 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-cni-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253637 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-netns\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253663 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-ovn\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253747 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-cni-multus\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253784 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bf580211-e03a-444e-98b8-c12f99091b80-cni-binary-copy\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253818 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/545e942f-5352-4a89-8e75-17762d512364-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253841 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-slash\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253902 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-config\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253946 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8490fb8-480e-4d81-82dd-4411b1f29278-ovn-node-metrics-cert\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.253987 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-mcd-auth-proxy-config\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.254013 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-var-lib-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.254059 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bf580211-e03a-444e-98b8-c12f99091b80-multus-daemon-config\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.254086 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bd7d\" (UniqueName: \"kubernetes.io/projected/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-kube-api-access-5bd7d\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.254110 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-system-cni-dir\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.265839 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.289637 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.313124 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.313797 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.313848 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.313942 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:34 crc kubenswrapper[4690]: E1203 13:09:34.314042 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.318832 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.319749 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.321497 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.322346 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.323773 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.324342 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.325031 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.326084 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.326770 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.328457 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.329023 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.330135 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.330717 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.331516 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.332495 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.333120 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.334448 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.334922 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.335555 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.336552 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.336775 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.337305 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.338368 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.338920 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.340012 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.340526 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.341205 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.342354 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.343083 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.344062 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.344585 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.345613 4690 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.345779 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.347443 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.348580 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.349108 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.351148 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.352244 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.353562 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354383 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354675 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-socket-dir-parent\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354726 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-conf-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354753 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-cnibin\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354787 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpngs\" (UniqueName: \"kubernetes.io/projected/bcf22e98-c7f6-4eb3-9c13-60f1953a9271-kube-api-access-tpngs\") pod \"node-resolver-qh4c5\" (UID: \"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\") " pod="openshift-dns/node-resolver-qh4c5" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354805 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354824 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-netns\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354841 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-ovn\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354863 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-cni-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354927 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-cni-multus\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354947 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bf580211-e03a-444e-98b8-c12f99091b80-cni-binary-copy\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354963 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/545e942f-5352-4a89-8e75-17762d512364-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.354982 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-slash\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355000 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-config\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355016 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8490fb8-480e-4d81-82dd-4411b1f29278-ovn-node-metrics-cert\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355039 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-mcd-auth-proxy-config\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355054 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-var-lib-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355075 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bf580211-e03a-444e-98b8-c12f99091b80-multus-daemon-config\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355091 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bd7d\" (UniqueName: \"kubernetes.io/projected/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-kube-api-access-5bd7d\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355107 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-system-cni-dir\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355129 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-kubelet\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355147 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-k8s-cni-cncf-io\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355164 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-systemd-units\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355201 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-netd\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355225 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-cni-bin\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355250 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-proxy-tls\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355273 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/545e942f-5352-4a89-8e75-17762d512364-cni-binary-copy\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355295 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-node-log\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355312 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-system-cni-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355325 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355636 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-socket-dir-parent\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355723 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-var-lib-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355749 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-os-release\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355773 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-netd\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355833 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-cni-bin\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355878 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-systemd-units\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355981 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-node-log\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355994 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-system-cni-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356041 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-system-cni-dir\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356030 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-cni-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356092 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-multus-conf-dir\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356175 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-cnibin\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356406 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-cni-multus\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356451 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-kubelet\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356582 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-k8s-cni-cncf-io\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356980 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/545e942f-5352-4a89-8e75-17762d512364-cni-binary-copy\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.356992 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357035 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-ovn\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357032 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bf580211-e03a-444e-98b8-c12f99091b80-multus-daemon-config\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.355330 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-os-release\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357216 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bf580211-e03a-444e-98b8-c12f99091b80-cni-binary-copy\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357267 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-netns\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357340 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357393 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-config\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357104 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-slash\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357517 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-etc-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357557 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-multus-certs\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357586 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-os-release\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357606 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-systemd\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357626 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-env-overrides\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357645 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-netns\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357665 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-hostroot\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357687 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-bin\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357706 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7hgf\" (UniqueName: \"kubernetes.io/projected/c8490fb8-480e-4d81-82dd-4411b1f29278-kube-api-access-r7hgf\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357730 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bcf22e98-c7f6-4eb3-9c13-60f1953a9271-hosts-file\") pod \"node-resolver-qh4c5\" (UID: \"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\") " pod="openshift-dns/node-resolver-qh4c5" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357750 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-rootfs\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357774 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357823 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwc65\" (UniqueName: \"kubernetes.io/projected/bf580211-e03a-444e-98b8-c12f99091b80-kube-api-access-nwc65\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357846 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357884 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-script-lib\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357923 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-cnibin\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357947 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-kubelet\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357970 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-etc-kubernetes\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357993 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxp5r\" (UniqueName: \"kubernetes.io/projected/545e942f-5352-4a89-8e75-17762d512364-kube-api-access-qxp5r\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358014 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-log-socket\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358039 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-ovn-kubernetes\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358109 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-ovn-kubernetes\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358165 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bcf22e98-c7f6-4eb3-9c13-60f1953a9271-hosts-file\") pod \"node-resolver-qh4c5\" (UID: \"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\") " pod="openshift-dns/node-resolver-qh4c5" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358200 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-rootfs\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.357749 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/545e942f-5352-4a89-8e75-17762d512364-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358251 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358309 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358400 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-netns\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358440 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-hostroot\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358452 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-env-overrides\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358463 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-bin\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358499 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-etc-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358530 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-systemd\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358531 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/545e942f-5352-4a89-8e75-17762d512364-os-release\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358554 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-run-multus-certs\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358556 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-openvswitch\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358583 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-host-var-lib-kubelet\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358602 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-etc-kubernetes\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358633 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bf580211-e03a-444e-98b8-c12f99091b80-cnibin\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.358663 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-log-socket\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.359085 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-script-lib\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.360342 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-mcd-auth-proxy-config\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.360643 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.361518 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.361916 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8490fb8-480e-4d81-82dd-4411b1f29278-ovn-node-metrics-cert\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.362007 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-proxy-tls\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.364312 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.365034 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.366076 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.367376 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.367939 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.368832 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.369333 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.370306 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.370931 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.371424 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.376085 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.385428 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwc65\" (UniqueName: \"kubernetes.io/projected/bf580211-e03a-444e-98b8-c12f99091b80-kube-api-access-nwc65\") pod \"multus-tf2cb\" (UID: \"bf580211-e03a-444e-98b8-c12f99091b80\") " pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.386125 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxp5r\" (UniqueName: \"kubernetes.io/projected/545e942f-5352-4a89-8e75-17762d512364-kube-api-access-qxp5r\") pod \"multus-additional-cni-plugins-dthlv\" (UID: \"545e942f-5352-4a89-8e75-17762d512364\") " pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.386289 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7hgf\" (UniqueName: \"kubernetes.io/projected/c8490fb8-480e-4d81-82dd-4411b1f29278-kube-api-access-r7hgf\") pod \"ovnkube-node-vmddn\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.387407 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpngs\" (UniqueName: \"kubernetes.io/projected/bcf22e98-c7f6-4eb3-9c13-60f1953a9271-kube-api-access-tpngs\") pod \"node-resolver-qh4c5\" (UID: \"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\") " pod="openshift-dns/node-resolver-qh4c5" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.401019 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.401759 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bd7d\" (UniqueName: \"kubernetes.io/projected/e23e6be9-62d7-4b0a-a073-52e9fc3a8723-kube-api-access-5bd7d\") pod \"machine-config-daemon-p5fc8\" (UID: \"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\") " pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.419175 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.432448 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.457022 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.459468 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qh4c5" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.467340 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:09:34 crc kubenswrapper[4690]: W1203 13:09:34.474582 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcf22e98_c7f6_4eb3_9c13_60f1953a9271.slice/crio-eaa8e19d8c53c8abf8c2509452de3951119ed4d90c34419d682d3f1b541e6afb WatchSource:0}: Error finding container eaa8e19d8c53c8abf8c2509452de3951119ed4d90c34419d682d3f1b541e6afb: Status 404 returned error can't find the container with id eaa8e19d8c53c8abf8c2509452de3951119ed4d90c34419d682d3f1b541e6afb Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.475386 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dthlv" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.476708 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.483661 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-tf2cb" Dec 03 13:09:34 crc kubenswrapper[4690]: W1203 13:09:34.484198 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode23e6be9_62d7_4b0a_a073_52e9fc3a8723.slice/crio-49f157ca107c1c307af9463e275a8517d7b1c5855d210d10920b81211ebbc375 WatchSource:0}: Error finding container 49f157ca107c1c307af9463e275a8517d7b1c5855d210d10920b81211ebbc375: Status 404 returned error can't find the container with id 49f157ca107c1c307af9463e275a8517d7b1c5855d210d10920b81211ebbc375 Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.492239 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.498850 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: W1203 13:09:34.516783 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf580211_e03a_444e_98b8_c12f99091b80.slice/crio-f59375526a71e469ec38a9282eab9b1945c735a47d96c938e76ff8a20da1c072 WatchSource:0}: Error finding container f59375526a71e469ec38a9282eab9b1945c735a47d96c938e76ff8a20da1c072: Status 404 returned error can't find the container with id f59375526a71e469ec38a9282eab9b1945c735a47d96c938e76ff8a20da1c072 Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.522295 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.537924 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.563126 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.585296 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.605468 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.620450 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.635029 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.649670 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.665694 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.679993 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.694467 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.734075 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.752229 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.770715 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.789542 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.812586 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.829181 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.845347 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.873051 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.898504 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.914772 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:34 crc kubenswrapper[4690]: I1203 13:09:34.945360 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.197263 4690 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.199250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.199284 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.199297 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.199396 4690 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.213783 4690 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.214405 4690 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.215834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.215859 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.215887 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.215906 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.215918 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.239406 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.246371 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.246441 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.246460 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.246485 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.246502 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.262463 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.266496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.266525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.266533 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.266548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.266560 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.280147 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.283635 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.283664 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.283674 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.283688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.283699 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.296892 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.300293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.300333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.300351 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.300369 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.300381 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.314760 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.314939 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.316058 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.316222 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.318203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.318240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.318254 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.318275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.318290 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.420102 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.420463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.420473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.420489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.420499 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.469384 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tf2cb" event={"ID":"bf580211-e03a-444e-98b8-c12f99091b80","Type":"ContainerStarted","Data":"7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.469531 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tf2cb" event={"ID":"bf580211-e03a-444e-98b8-c12f99091b80","Type":"ContainerStarted","Data":"f59375526a71e469ec38a9282eab9b1945c735a47d96c938e76ff8a20da1c072"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.480565 4690 generic.go:334] "Generic (PLEG): container finished" podID="545e942f-5352-4a89-8e75-17762d512364" containerID="d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba" exitCode=0 Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.480751 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" event={"ID":"545e942f-5352-4a89-8e75-17762d512364","Type":"ContainerDied","Data":"d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.480814 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" event={"ID":"545e942f-5352-4a89-8e75-17762d512364","Type":"ContainerStarted","Data":"b15d660dea706473ca5048bd68553100b6291d76b1d760bd9931b3835966db47"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.490225 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.500603 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.500940 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.501019 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.501038 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"49f157ca107c1c307af9463e275a8517d7b1c5855d210d10920b81211ebbc375"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.503168 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qh4c5" event={"ID":"bcf22e98-c7f6-4eb3-9c13-60f1953a9271","Type":"ContainerStarted","Data":"9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.503234 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qh4c5" event={"ID":"bcf22e98-c7f6-4eb3-9c13-60f1953a9271","Type":"ContainerStarted","Data":"eaa8e19d8c53c8abf8c2509452de3951119ed4d90c34419d682d3f1b541e6afb"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.504528 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647" exitCode=0 Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.504572 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.504594 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"6c2a01d440258f2faa2abbb255797674c12b4118b69c7fc65a55cf7500b4df35"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.521723 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.526225 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.526272 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.526289 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.526308 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.526323 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.539718 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.556902 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.571857 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.590506 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.606992 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.629796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.629840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.629851 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.629900 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.629913 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.632688 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.647541 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.664818 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.680778 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.701293 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.718468 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.731646 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.733640 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.733671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.733681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.733697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.733711 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.747016 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.766104 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.786070 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.807515 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.824140 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.836492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.836526 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.836535 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.836553 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.836565 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.842619 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.859220 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.881409 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.897745 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.915272 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.929026 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.941728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.941772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.941785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.941802 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.941812 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:35Z","lastTransitionTime":"2025-12-03T13:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.943370 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.954897 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.969785 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.977699 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.977815 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:35 crc kubenswrapper[4690]: I1203 13:09:35.977844 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.977981 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:09:39.977938485 +0000 UTC m=+25.958858918 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.978002 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.978109 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:39.978086098 +0000 UTC m=+25.959006531 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.978033 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:35 crc kubenswrapper[4690]: E1203 13:09:35.978157 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:39.97815013 +0000 UTC m=+25.959070563 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.044142 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.044171 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.044180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.044198 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.044208 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.078387 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.078440 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.078587 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.078607 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.078620 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.078673 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:40.078656514 +0000 UTC m=+26.059576947 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.079087 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.079111 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.079122 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.079152 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:40.079142916 +0000 UTC m=+26.060063349 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.146068 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.146098 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.146106 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.146120 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.146129 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.248727 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.248767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.248776 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.248792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.248807 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.314476 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.314563 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.314627 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:36 crc kubenswrapper[4690]: E1203 13:09:36.314691 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.351226 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.351320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.351340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.351359 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.351370 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.453783 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.453829 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.453841 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.453859 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.453894 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.512131 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.512191 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.512205 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.512217 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.512229 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.512240 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.513578 4690 generic.go:334] "Generic (PLEG): container finished" podID="545e942f-5352-4a89-8e75-17762d512364" containerID="a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1" exitCode=0 Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.513970 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" event={"ID":"545e942f-5352-4a89-8e75-17762d512364","Type":"ContainerDied","Data":"a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.527885 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.549287 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.556570 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.556605 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.556617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.556634 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.556644 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.564242 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.577275 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.596476 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.619254 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.634046 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.645594 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.656471 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.660496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.660755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.660768 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.660805 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.660849 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.670452 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.687610 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.704148 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.717791 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.735961 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:36Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.764130 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.764178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.764188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.764208 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.764219 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.867027 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.867089 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.867101 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.867123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.867134 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.969041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.969088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.969103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.969130 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:36 crc kubenswrapper[4690]: I1203 13:09:36.969143 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:36Z","lastTransitionTime":"2025-12-03T13:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.072135 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.072178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.072188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.072212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.072229 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.175354 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.175399 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.175412 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.175428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.175439 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.278243 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.278279 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.278288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.278306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.278327 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.313937 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:37 crc kubenswrapper[4690]: E1203 13:09:37.314074 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.381001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.381033 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.381042 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.381055 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.381065 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.483737 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.483811 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.483825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.483849 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.483861 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.518762 4690 generic.go:334] "Generic (PLEG): container finished" podID="545e942f-5352-4a89-8e75-17762d512364" containerID="8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf" exitCode=0 Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.518807 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" event={"ID":"545e942f-5352-4a89-8e75-17762d512364","Type":"ContainerDied","Data":"8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.532288 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.543653 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.556766 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.574820 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.586275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.586314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.586326 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.586342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.586354 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.589061 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.600648 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.613815 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.630154 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.654590 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.678143 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.689823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.689877 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.689890 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.689908 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.689925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.701827 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.720111 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.736166 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.749376 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.793369 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.793405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.793416 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.793434 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.793446 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.895619 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.895651 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.895662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.895677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.895689 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.997773 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.997812 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.997821 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.997835 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:37 crc kubenswrapper[4690]: I1203 13:09:37.997847 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:37Z","lastTransitionTime":"2025-12-03T13:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.020564 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-gkl8h"] Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.020971 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.023023 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.023559 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.023713 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.024474 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.044726 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.065615 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.083695 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.096206 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f190c94-af05-4d67-8459-c02f4ecc81b3-host\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.096273 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f190c94-af05-4d67-8459-c02f4ecc81b3-serviceca\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.096340 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hpd2\" (UniqueName: \"kubernetes.io/projected/4f190c94-af05-4d67-8459-c02f4ecc81b3-kube-api-access-2hpd2\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.099297 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.100883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.100921 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.100934 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.100952 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.100966 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.112525 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.147733 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.165186 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.197302 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f190c94-af05-4d67-8459-c02f4ecc81b3-host\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.197380 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f190c94-af05-4d67-8459-c02f4ecc81b3-serviceca\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.197408 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hpd2\" (UniqueName: \"kubernetes.io/projected/4f190c94-af05-4d67-8459-c02f4ecc81b3-kube-api-access-2hpd2\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.197465 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f190c94-af05-4d67-8459-c02f4ecc81b3-host\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.198908 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f190c94-af05-4d67-8459-c02f4ecc81b3-serviceca\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.203851 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.203915 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.203931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.203954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.203968 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.214615 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.223467 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hpd2\" (UniqueName: \"kubernetes.io/projected/4f190c94-af05-4d67-8459-c02f4ecc81b3-kube-api-access-2hpd2\") pod \"node-ca-gkl8h\" (UID: \"4f190c94-af05-4d67-8459-c02f4ecc81b3\") " pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.231992 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.244763 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.259325 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.275120 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.293084 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.306635 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.306679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.306692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.306709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.306723 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.308843 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.314558 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.314622 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:38 crc kubenswrapper[4690]: E1203 13:09:38.314720 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:38 crc kubenswrapper[4690]: E1203 13:09:38.314790 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.325546 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.334617 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gkl8h" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.411152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.411565 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.411578 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.411594 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.411605 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.514638 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.514682 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.514695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.514717 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.514728 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.530658 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.533664 4690 generic.go:334] "Generic (PLEG): container finished" podID="545e942f-5352-4a89-8e75-17762d512364" containerID="76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0" exitCode=0 Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.533720 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" event={"ID":"545e942f-5352-4a89-8e75-17762d512364","Type":"ContainerDied","Data":"76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.536032 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gkl8h" event={"ID":"4f190c94-af05-4d67-8459-c02f4ecc81b3","Type":"ContainerStarted","Data":"72626b84733f603e3368885f021f4cd95a0e975af1dd2e9d858d4d969bdb7b8a"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.549980 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.566910 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.582771 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.596918 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.614657 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.617674 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.617718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.617732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.617751 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.617763 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.628780 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.646347 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.667365 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.686334 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.703134 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.715663 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.720316 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.720370 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.720381 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.720400 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.720414 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.735085 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.746271 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.758160 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.772362 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.823806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.823846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.823856 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.823886 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.823897 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.926460 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.926522 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.926535 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.926554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:38 crc kubenswrapper[4690]: I1203 13:09:38.926568 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:38Z","lastTransitionTime":"2025-12-03T13:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.028729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.028771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.028782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.028798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.028811 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.131048 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.131078 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.131086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.131099 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.131107 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.233958 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.234025 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.234049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.234079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.234101 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.314396 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:39 crc kubenswrapper[4690]: E1203 13:09:39.314569 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.336920 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.336970 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.336984 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.337005 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.337022 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.439421 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.439463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.439472 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.439488 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.439498 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.541217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.541575 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.541586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.541601 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.541611 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.542949 4690 generic.go:334] "Generic (PLEG): container finished" podID="545e942f-5352-4a89-8e75-17762d512364" containerID="a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c" exitCode=0 Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.542996 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" event={"ID":"545e942f-5352-4a89-8e75-17762d512364","Type":"ContainerDied","Data":"a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.545390 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gkl8h" event={"ID":"4f190c94-af05-4d67-8459-c02f4ecc81b3","Type":"ContainerStarted","Data":"2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.556050 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.567200 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.580545 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.595115 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.609095 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.623020 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.636531 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.647545 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.647584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.647594 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.647609 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.647620 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.648746 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.661390 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.678123 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.690377 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.702970 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.724607 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.734815 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.751345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.751386 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.751400 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.751417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.751429 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.754185 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.766038 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.776972 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.790624 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.803155 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.814496 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.826412 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.837689 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.852350 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.853532 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.853558 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.853569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.853584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.853594 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.868843 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.891897 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.904574 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.917007 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.928271 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.955917 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.955963 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.955973 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.955990 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.956002 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:39Z","lastTransitionTime":"2025-12-03T13:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.971695 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:39 crc kubenswrapper[4690]: I1203 13:09:39.991774 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.012802 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.013031 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:09:48.013000898 +0000 UTC m=+33.993921331 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.013262 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.013418 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.013462 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:48.013454129 +0000 UTC m=+33.994374552 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.013562 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.013409 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.013652 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:48.013629053 +0000 UTC m=+33.994549546 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.058340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.058376 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.058386 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.058400 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.058409 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.115175 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.115231 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.115372 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.115394 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.115407 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.115458 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:48.11544172 +0000 UTC m=+34.096362153 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.115477 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.115561 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.115582 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.115686 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:09:48.115652425 +0000 UTC m=+34.096572868 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.161348 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.161636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.161730 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.161817 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.161924 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.264631 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.264668 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.264681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.264699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.264711 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.314660 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.314708 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.314824 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:40 crc kubenswrapper[4690]: E1203 13:09:40.315060 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.367448 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.367494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.367507 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.367524 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.367533 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.470606 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.470660 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.470677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.470696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.470713 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.553815 4690 generic.go:334] "Generic (PLEG): container finished" podID="545e942f-5352-4a89-8e75-17762d512364" containerID="438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928" exitCode=0 Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.554398 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" event={"ID":"545e942f-5352-4a89-8e75-17762d512364","Type":"ContainerDied","Data":"438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.569441 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.574004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.574047 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.574056 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.574071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.574084 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.583989 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.600349 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.615494 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.629771 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.646464 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.661470 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.679312 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.679360 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.679371 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.679391 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.679404 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.680048 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.694277 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.712837 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.731148 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.742627 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.752792 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.773391 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.781617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.781648 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.781658 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.781674 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.781687 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.784978 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.884619 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.884671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.884688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.884709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.884725 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.987019 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.987054 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.987062 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.987076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:40 crc kubenswrapper[4690]: I1203 13:09:40.987086 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:40Z","lastTransitionTime":"2025-12-03T13:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.088838 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.088885 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.088896 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.088910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.088918 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.191246 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.191270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.191279 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.191293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.191302 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.294102 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.294133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.294141 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.294156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.294165 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.314513 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:41 crc kubenswrapper[4690]: E1203 13:09:41.314754 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.396239 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.396294 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.396306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.396329 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.396345 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.499151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.499589 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.499706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.499805 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.499889 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.562393 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.563039 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.568476 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" event={"ID":"545e942f-5352-4a89-8e75-17762d512364","Type":"ContainerStarted","Data":"398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.579673 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.628357 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.628419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.628434 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.628457 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.628473 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.631637 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.634066 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.661527 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.675388 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.704027 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.719493 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.731116 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.731155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.731167 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.731182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.731195 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.733180 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.749806 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.766001 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.782236 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.800579 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.818031 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.835147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.835188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.835200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.835215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.835226 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.835322 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.856345 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.874312 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.888844 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.903338 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.921370 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.937715 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.938447 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.938551 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.938566 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.938643 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.938667 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:41Z","lastTransitionTime":"2025-12-03T13:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.952166 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.977537 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:41 crc kubenswrapper[4690]: I1203 13:09:41.991015 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:41Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.015453 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.033563 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.041479 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.041535 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.041544 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.041562 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.041573 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.047372 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.061836 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.074503 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.088938 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.104819 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.117522 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.144240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.144306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.144316 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.144334 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.144345 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.246624 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.246670 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.246681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.246697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.246710 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.314703 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.314716 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:42 crc kubenswrapper[4690]: E1203 13:09:42.314978 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:42 crc kubenswrapper[4690]: E1203 13:09:42.315125 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.349319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.349389 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.349404 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.349425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.349437 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.452957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.453009 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.453020 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.453039 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.453054 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.557179 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.557225 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.557234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.557252 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.557267 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.576509 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.578692 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.616437 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.633547 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.647064 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.660117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.660223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.660238 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.660256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.660267 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.661954 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.675171 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.690668 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.705409 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.718573 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.734879 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.748618 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.761671 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.765081 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.765124 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.765135 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.765152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.765166 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.786060 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.798077 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.819050 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.824718 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.831513 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.842868 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.868425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.868471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.868481 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.868500 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.868512 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.868762 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.881154 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.894021 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.909117 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.927753 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.939519 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.953395 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.965628 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.971160 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.971206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.971218 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.971236 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.971248 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:42Z","lastTransitionTime":"2025-12-03T13:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:42 crc kubenswrapper[4690]: I1203 13:09:42.981891 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.000116 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:42Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.015540 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.030683 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.045469 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.058751 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.070791 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.073683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.073721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.073732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.073749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.073764 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.175906 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.175948 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.175959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.175976 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.175986 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.282890 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.282945 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.282957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.282979 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.282997 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.314016 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:43 crc kubenswrapper[4690]: E1203 13:09:43.314324 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.385586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.385638 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.385648 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.385667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.385679 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.488559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.488617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.488636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.488667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.488688 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.577889 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.592728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.592946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.592965 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.592989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.593009 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.696510 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.696582 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.696602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.696629 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.696649 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.800121 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.800177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.800191 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.800211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.800224 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.903193 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.903250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.903264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.903284 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:43 crc kubenswrapper[4690]: I1203 13:09:43.903298 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:43Z","lastTransitionTime":"2025-12-03T13:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.006757 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.006848 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.006861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.006888 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.006915 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.109459 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.109718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.109820 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.109918 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.110019 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.212131 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.212439 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.212508 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.212589 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.212654 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.314103 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:44 crc kubenswrapper[4690]: E1203 13:09:44.314370 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.314569 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:44 crc kubenswrapper[4690]: E1203 13:09:44.314792 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.315312 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.315446 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.315542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.315647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.315738 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.332391 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.349361 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.365912 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.380515 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.394173 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.409895 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.418356 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.418404 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.418415 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.418435 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.418448 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.429023 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.440206 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.459974 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.475842 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.488995 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.504494 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.518146 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.520806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.520845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.520858 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.520901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.520925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.533994 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.550797 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.583431 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/0.log" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.586842 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63" exitCode=1 Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.586889 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.588140 4690 scope.go:117] "RemoveContainer" containerID="14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.601431 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.622540 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:43Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:43.102092 5966 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:09:43.102115 5966 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:09:43.102163 5966 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:09:43.102190 5966 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 13:09:43.102213 5966 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 13:09:43.102224 5966 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:09:43.102235 5966 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:43.102311 5966 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:43.102312 5966 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 13:09:43.102344 5966 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:43.102351 5966 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:43.102386 5966 factory.go:656] Stopping watch factory\\\\nI1203 13:09:43.102405 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:43.102431 5966 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:43.102439 5966 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.624942 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.624992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.625007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.625026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.625039 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.636656 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.659976 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.675697 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.690146 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.704018 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.715817 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.727909 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.727960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.727972 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.727996 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.728010 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.729899 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.742574 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.761108 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.773952 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.785666 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.800509 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.813339 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.830997 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.831045 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.831059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.831076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.831088 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.933122 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.933161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.933170 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.933187 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:44 crc kubenswrapper[4690]: I1203 13:09:44.933199 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:44Z","lastTransitionTime":"2025-12-03T13:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.035636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.035657 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.035667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.035680 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.035690 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.138234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.138270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.138315 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.138329 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.138338 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.241336 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.241402 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.241421 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.241454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.241475 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.254501 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.314837 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:45 crc kubenswrapper[4690]: E1203 13:09:45.315084 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.345137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.345186 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.345203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.345224 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.345239 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.449587 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.449648 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.449661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.449683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.449695 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.479847 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.479998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.480029 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.480070 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.480104 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: E1203 13:09:45.501229 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.506160 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.506276 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.506328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.506361 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.506408 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: E1203 13:09:45.523448 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.527673 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.527721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.527732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.527752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.527793 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: E1203 13:09:45.540235 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.544384 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.544435 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.544452 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.544474 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.544490 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: E1203 13:09:45.559847 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.563715 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.563750 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.563761 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.563778 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.563790 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: E1203 13:09:45.589028 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: E1203 13:09:45.589187 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.592012 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.592073 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.592085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.592101 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.592119 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/0.log" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.592114 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.596143 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.596917 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.617935 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.640041 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.653165 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.675665 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.689097 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.695076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.695113 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.695127 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.695144 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.695158 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.704040 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.716401 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.735219 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:43Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:43.102092 5966 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:09:43.102115 5966 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:09:43.102163 5966 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:09:43.102190 5966 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 13:09:43.102213 5966 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 13:09:43.102224 5966 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:09:43.102235 5966 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:43.102311 5966 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:43.102312 5966 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 13:09:43.102344 5966 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:43.102351 5966 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:43.102386 5966 factory.go:656] Stopping watch factory\\\\nI1203 13:09:43.102405 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:43.102431 5966 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:43.102439 5966 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.746339 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.758721 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.769549 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.785865 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.797464 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.797676 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.797693 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.797701 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.797713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.797722 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.813817 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.826781 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:45Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.900358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.900453 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.900471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.900497 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:45 crc kubenswrapper[4690]: I1203 13:09:45.900537 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:45Z","lastTransitionTime":"2025-12-03T13:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.003353 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.003433 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.003450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.003475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.003515 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.106018 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.106063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.106075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.106094 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.106105 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.162007 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d"] Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.162540 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.164574 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.165581 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.181752 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.189433 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk5h9\" (UniqueName: \"kubernetes.io/projected/7f1e382c-408f-4b5c-a51f-2f90770af7fe-kube-api-access-hk5h9\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.189478 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7f1e382c-408f-4b5c-a51f-2f90770af7fe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.189533 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7f1e382c-408f-4b5c-a51f-2f90770af7fe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.189555 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7f1e382c-408f-4b5c-a51f-2f90770af7fe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.195305 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.206635 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.208074 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.208107 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.208119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.208157 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.208169 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.219257 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.230692 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.246562 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.259495 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.270638 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.281606 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.290278 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7f1e382c-408f-4b5c-a51f-2f90770af7fe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.290318 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk5h9\" (UniqueName: \"kubernetes.io/projected/7f1e382c-408f-4b5c-a51f-2f90770af7fe-kube-api-access-hk5h9\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.290336 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7f1e382c-408f-4b5c-a51f-2f90770af7fe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.290391 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7f1e382c-408f-4b5c-a51f-2f90770af7fe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.291005 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7f1e382c-408f-4b5c-a51f-2f90770af7fe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.291233 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7f1e382c-408f-4b5c-a51f-2f90770af7fe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.296232 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7f1e382c-408f-4b5c-a51f-2f90770af7fe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.301121 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.308572 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk5h9\" (UniqueName: \"kubernetes.io/projected/7f1e382c-408f-4b5c-a51f-2f90770af7fe-kube-api-access-hk5h9\") pod \"ovnkube-control-plane-749d76644c-2hc7d\" (UID: \"7f1e382c-408f-4b5c-a51f-2f90770af7fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.310782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.310825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.310838 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.310855 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.310869 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.314212 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.314326 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:46 crc kubenswrapper[4690]: E1203 13:09:46.314418 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:46 crc kubenswrapper[4690]: E1203 13:09:46.314519 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.315837 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.329595 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.343546 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.366658 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:43Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:43.102092 5966 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:09:43.102115 5966 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:09:43.102163 5966 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:09:43.102190 5966 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 13:09:43.102213 5966 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 13:09:43.102224 5966 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:09:43.102235 5966 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:43.102311 5966 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:43.102312 5966 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 13:09:43.102344 5966 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:43.102351 5966 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:43.102386 5966 factory.go:656] Stopping watch factory\\\\nI1203 13:09:43.102405 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:43.102431 5966 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:43.102439 5966 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.380932 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.392489 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:46Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.413887 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.413939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.413949 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.413964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.413976 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.476591 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" Dec 03 13:09:46 crc kubenswrapper[4690]: W1203 13:09:46.493188 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f1e382c_408f_4b5c_a51f_2f90770af7fe.slice/crio-03ef7032ce8707c883cb6f9a442980be78f651b40bccf3b4e17e1f6898cba4a1 WatchSource:0}: Error finding container 03ef7032ce8707c883cb6f9a442980be78f651b40bccf3b4e17e1f6898cba4a1: Status 404 returned error can't find the container with id 03ef7032ce8707c883cb6f9a442980be78f651b40bccf3b4e17e1f6898cba4a1 Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.516747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.516813 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.516829 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.516852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.516900 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.600580 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" event={"ID":"7f1e382c-408f-4b5c-a51f-2f90770af7fe","Type":"ContainerStarted","Data":"03ef7032ce8707c883cb6f9a442980be78f651b40bccf3b4e17e1f6898cba4a1"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.620230 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.620281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.620297 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.620318 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.620332 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.722496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.722562 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.722574 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.722622 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.722635 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.833713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.833765 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.833781 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.833802 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.833818 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.937756 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.937839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.937853 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.937877 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:46 crc kubenswrapper[4690]: I1203 13:09:46.937907 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:46Z","lastTransitionTime":"2025-12-03T13:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.040946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.040985 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.040994 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.041011 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.041022 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.144511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.145001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.145014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.145030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.145042 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.268441 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.268479 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.268489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.268503 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.268512 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.314123 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:47 crc kubenswrapper[4690]: E1203 13:09:47.314244 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.370723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.370759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.370771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.370790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.370801 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.473380 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.473428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.473439 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.473456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.473467 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.575660 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.575727 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.575752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.575782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.575805 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.607837 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/1.log" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.609049 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/0.log" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.613053 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06" exitCode=1 Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.613145 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.613204 4690 scope.go:117] "RemoveContainer" containerID="14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.615196 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" event={"ID":"7f1e382c-408f-4b5c-a51f-2f90770af7fe","Type":"ContainerStarted","Data":"affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.615257 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" event={"ID":"7f1e382c-408f-4b5c-a51f-2f90770af7fe","Type":"ContainerStarted","Data":"66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.615409 4690 scope.go:117] "RemoveContainer" containerID="e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06" Dec 03 13:09:47 crc kubenswrapper[4690]: E1203 13:09:47.615808 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.636087 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.653045 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.668481 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.678688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.678729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.678741 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.678760 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.678773 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.680404 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.690469 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.701182 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.712987 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.727218 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.739427 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.748997 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.771953 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.782106 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.782135 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.782147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.782164 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.782175 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.783774 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.794520 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.804121 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.821003 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:43Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:43.102092 5966 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:09:43.102115 5966 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:09:43.102163 5966 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:09:43.102190 5966 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 13:09:43.102213 5966 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 13:09:43.102224 5966 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:09:43.102235 5966 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:43.102311 5966 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:43.102312 5966 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 13:09:43.102344 5966 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:43.102351 5966 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:43.102386 5966 factory.go:656] Stopping watch factory\\\\nI1203 13:09:43.102405 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:43.102431 5966 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:43.102439 5966 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"message\\\":\\\"d\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 13:09:47.297861 6116 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:09:47.297958 6116 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-webhook per-node LB for network=default: []services.LB{}\\\\nF1203 13:09:47.297967 6116 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.830814 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.840395 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.852343 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.862735 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.872926 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.881213 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.884758 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.884785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.884805 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.884824 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.884833 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.901501 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.913821 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.925466 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.935227 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.954928 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:43Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:43.102092 5966 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:09:43.102115 5966 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:09:43.102163 5966 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:09:43.102190 5966 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 13:09:43.102213 5966 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 13:09:43.102224 5966 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:09:43.102235 5966 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:43.102311 5966 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:43.102312 5966 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 13:09:43.102344 5966 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:43.102351 5966 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:43.102386 5966 factory.go:656] Stopping watch factory\\\\nI1203 13:09:43.102405 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:43.102431 5966 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:43.102439 5966 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"message\\\":\\\"d\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 13:09:47.297861 6116 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:09:47.297958 6116 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-webhook per-node LB for network=default: []services.LB{}\\\\nF1203 13:09:47.297967 6116 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.967992 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.975786 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.986273 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.987026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.987059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.987067 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.987083 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.987093 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:47Z","lastTransitionTime":"2025-12-03T13:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:47 crc kubenswrapper[4690]: I1203 13:09:47.996885 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.007993 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.020260 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-mbbpr"] Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.020953 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.021651 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.021852 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.032578 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.042365 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.054108 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.066379 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.076811 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.090320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.090692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.090755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.090819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.090904 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.091603 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.104036 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.109533 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.109812 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:10:04.109774748 +0000 UTC m=+50.090695181 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.109901 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.109992 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p92gs\" (UniqueName: \"kubernetes.io/projected/abab6c63-91d1-4f36-9be9-ae3b1a43742e-kube-api-access-p92gs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.110034 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.110074 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.110286 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.110350 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:10:04.110338122 +0000 UTC m=+50.091258735 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.110430 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.110557 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:10:04.110536297 +0000 UTC m=+50.091456730 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.115967 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.128064 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.138941 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.148961 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.158806 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.175073 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.193513 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.193869 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.193918 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.193931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.193952 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.193966 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.205691 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.211427 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p92gs\" (UniqueName: \"kubernetes.io/projected/abab6c63-91d1-4f36-9be9-ae3b1a43742e-kube-api-access-p92gs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.211616 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.211671 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211717 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211739 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211752 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.211748 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211799 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:10:04.211784739 +0000 UTC m=+50.192705162 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211849 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211882 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211895 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211942 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.211945 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:10:04.211927263 +0000 UTC m=+50.192847696 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.212036 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs podName:abab6c63-91d1-4f36-9be9-ae3b1a43742e nodeName:}" failed. No retries permitted until 2025-12-03 13:09:48.712012035 +0000 UTC m=+34.692932508 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs") pod "network-metrics-daemon-mbbpr" (UID: "abab6c63-91d1-4f36-9be9-ae3b1a43742e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.219733 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.229924 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p92gs\" (UniqueName: \"kubernetes.io/projected/abab6c63-91d1-4f36-9be9-ae3b1a43742e-kube-api-access-p92gs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.241720 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:43Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:43.102092 5966 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:09:43.102115 5966 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:09:43.102163 5966 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:09:43.102190 5966 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 13:09:43.102213 5966 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 13:09:43.102224 5966 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:09:43.102235 5966 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:43.102311 5966 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:43.102312 5966 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 13:09:43.102344 5966 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:43.102351 5966 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:43.102386 5966 factory.go:656] Stopping watch factory\\\\nI1203 13:09:43.102405 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:43.102431 5966 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:43.102439 5966 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"message\\\":\\\"d\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 13:09:47.297861 6116 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:09:47.297958 6116 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-webhook per-node LB for network=default: []services.LB{}\\\\nF1203 13:09:47.297967 6116 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:48Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.296193 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.296227 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.296238 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.296255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.296268 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.313817 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.313957 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.314064 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.314175 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.398620 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.398696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.398720 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.398748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.398766 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.501962 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.502008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.502019 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.502034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.502043 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.604748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.604791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.604804 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.604825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.604840 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.622076 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/1.log" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.707083 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.707134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.707145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.707162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.707176 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.717802 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.717988 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: E1203 13:09:48.718060 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs podName:abab6c63-91d1-4f36-9be9-ae3b1a43742e nodeName:}" failed. No retries permitted until 2025-12-03 13:09:49.718041542 +0000 UTC m=+35.698961975 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs") pod "network-metrics-daemon-mbbpr" (UID: "abab6c63-91d1-4f36-9be9-ae3b1a43742e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.809411 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.809461 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.809473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.809491 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.809502 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.911423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.911486 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.911502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.911521 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:48 crc kubenswrapper[4690]: I1203 13:09:48.911535 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:48Z","lastTransitionTime":"2025-12-03T13:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.013939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.013986 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.013995 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.014011 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.014022 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.116119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.116162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.116172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.116188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.116200 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.218604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.218643 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.218660 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.218677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.218689 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.314127 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.314141 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:49 crc kubenswrapper[4690]: E1203 13:09:49.314295 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:49 crc kubenswrapper[4690]: E1203 13:09:49.314360 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.321254 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.321314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.321326 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.321342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.321352 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.424504 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.424579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.424604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.424635 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.424658 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.527182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.527229 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.527243 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.527262 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.527275 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.629077 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.629145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.629164 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.629189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.629207 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.727531 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:49 crc kubenswrapper[4690]: E1203 13:09:49.728009 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:49 crc kubenswrapper[4690]: E1203 13:09:49.728149 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs podName:abab6c63-91d1-4f36-9be9-ae3b1a43742e nodeName:}" failed. No retries permitted until 2025-12-03 13:09:51.728128666 +0000 UTC m=+37.709049099 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs") pod "network-metrics-daemon-mbbpr" (UID: "abab6c63-91d1-4f36-9be9-ae3b1a43742e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.732304 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.732358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.732377 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.732396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.732408 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.835549 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.835596 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.835606 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.835620 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.835632 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.938335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.938662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.938817 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.939218 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:49 crc kubenswrapper[4690]: I1203 13:09:49.939605 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:49Z","lastTransitionTime":"2025-12-03T13:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.041779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.041821 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.041831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.041850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.041865 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.144313 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.144348 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.144358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.144373 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.144384 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.247517 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.247570 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.247582 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.247601 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.247613 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.314769 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:50 crc kubenswrapper[4690]: E1203 13:09:50.314909 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.314774 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:50 crc kubenswrapper[4690]: E1203 13:09:50.315006 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.350651 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.351049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.351293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.351551 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.351810 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.455055 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.455114 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.455133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.455162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.455179 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.559524 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.559606 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.559676 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.559708 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.559725 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.663132 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.664550 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.664598 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.664620 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.664641 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.767147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.767204 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.767221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.767243 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.767261 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.870346 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.870396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.870414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.870439 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.870462 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.973217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.973277 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.973294 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.973326 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:50 crc kubenswrapper[4690]: I1203 13:09:50.973343 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:50Z","lastTransitionTime":"2025-12-03T13:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.076922 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.077205 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.077296 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.077385 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.077501 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.180105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.180426 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.180561 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.180690 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.180816 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.283776 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.283816 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.283827 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.283843 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.283855 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.313936 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:51 crc kubenswrapper[4690]: E1203 13:09:51.314054 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.314465 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:51 crc kubenswrapper[4690]: E1203 13:09:51.314685 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.386308 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.386376 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.386395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.386421 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.386439 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.489637 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.489693 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.489707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.489730 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.489743 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.592513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.592567 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.592577 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.592598 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.592610 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.695299 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.695362 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.695371 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.695389 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.695399 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.748612 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:51 crc kubenswrapper[4690]: E1203 13:09:51.748820 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:51 crc kubenswrapper[4690]: E1203 13:09:51.748903 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs podName:abab6c63-91d1-4f36-9be9-ae3b1a43742e nodeName:}" failed. No retries permitted until 2025-12-03 13:09:55.748856817 +0000 UTC m=+41.729777250 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs") pod "network-metrics-daemon-mbbpr" (UID: "abab6c63-91d1-4f36-9be9-ae3b1a43742e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.798536 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.798602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.798613 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.798628 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.798637 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.901899 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.901948 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.901991 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.902008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:51 crc kubenswrapper[4690]: I1203 13:09:51.902019 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:51Z","lastTransitionTime":"2025-12-03T13:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.004512 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.004557 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.004566 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.004581 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.004593 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.107825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.107927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.107956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.107987 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.108008 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.210101 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.210138 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.210147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.210164 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.210178 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.317004 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:52 crc kubenswrapper[4690]: E1203 13:09:52.317585 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.317101 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:52 crc kubenswrapper[4690]: E1203 13:09:52.317731 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.320664 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.320797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.320894 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.320964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.321023 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.423565 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.423618 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.423631 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.423655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.423669 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.527146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.527189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.527198 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.527214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.527225 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.630175 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.630257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.630272 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.630293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.630311 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.733476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.734155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.734220 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.734247 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.734270 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.837201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.837291 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.837310 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.837335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.837353 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.939977 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.940017 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.940034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.940049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:52 crc kubenswrapper[4690]: I1203 13:09:52.940059 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:52Z","lastTransitionTime":"2025-12-03T13:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.042523 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.042554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.042563 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.042576 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.042585 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.145131 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.145170 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.145179 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.145193 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.145201 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.247733 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.247777 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.247787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.247802 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.247818 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.314622 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.314688 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:53 crc kubenswrapper[4690]: E1203 13:09:53.314767 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:53 crc kubenswrapper[4690]: E1203 13:09:53.314836 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.349978 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.350035 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.350046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.350063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.350090 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.452376 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.452448 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.452462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.452480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.452496 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.555407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.555463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.555476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.555500 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.555513 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.658008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.658059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.658068 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.658084 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.658098 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.760285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.760344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.760357 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.760377 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.760390 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.862602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.862654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.862663 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.862678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.862688 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.965369 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.965423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.965438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.965457 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:53 crc kubenswrapper[4690]: I1203 13:09:53.965469 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:53Z","lastTransitionTime":"2025-12-03T13:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.068383 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.068774 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.068850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.068957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.069039 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.172061 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.172496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.172560 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.172640 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.172713 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.275221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.275524 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.275596 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.275661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.275727 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.314483 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.314594 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:54 crc kubenswrapper[4690]: E1203 13:09:54.314606 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:54 crc kubenswrapper[4690]: E1203 13:09:54.314802 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.332725 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.347749 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.362855 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.377264 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.378711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.378738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.378749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.378767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.378779 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.390223 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.406575 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.424289 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.439305 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.452114 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.475221 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.480756 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.481085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.481224 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.481355 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.481476 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.490051 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.503221 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.516246 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.534758 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ff97be7e54beaf0171c3509d05e93305131abaf97bec922d647f5ab1f8fb63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:43Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:43.102092 5966 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:09:43.102115 5966 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:09:43.102163 5966 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:09:43.102190 5966 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 13:09:43.102213 5966 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 13:09:43.102224 5966 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:09:43.102235 5966 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:43.102311 5966 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:43.102312 5966 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 13:09:43.102344 5966 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:43.102351 5966 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:43.102386 5966 factory.go:656] Stopping watch factory\\\\nI1203 13:09:43.102405 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:43.102431 5966 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:43.102439 5966 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"message\\\":\\\"d\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 13:09:47.297861 6116 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:09:47.297958 6116 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-webhook per-node LB for network=default: []services.LB{}\\\\nF1203 13:09:47.297967 6116 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.546051 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.558006 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.570732 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.583776 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.583829 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.583847 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.583902 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.583921 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.686450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.686495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.686530 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.686552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.686569 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.788811 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.788851 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.788884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.788902 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.788913 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.891021 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.891089 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.891106 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.891133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.891152 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.994062 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.994096 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.994105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.994119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:54 crc kubenswrapper[4690]: I1203 13:09:54.994129 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:54Z","lastTransitionTime":"2025-12-03T13:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.096445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.096488 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.096502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.096518 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.096530 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.199128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.199161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.199170 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.199184 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.199194 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.301684 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.301753 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.301766 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.301788 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.301802 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.314099 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.314099 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:55 crc kubenswrapper[4690]: E1203 13:09:55.314301 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:09:55 crc kubenswrapper[4690]: E1203 13:09:55.314421 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.404550 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.404682 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.404698 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.404723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.404736 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.507639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.507685 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.507698 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.507717 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.507726 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.610823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.610881 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.610891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.610908 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.610918 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.713529 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.713590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.713607 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.713636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.713657 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.797990 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:55 crc kubenswrapper[4690]: E1203 13:09:55.798241 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:55 crc kubenswrapper[4690]: E1203 13:09:55.798349 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs podName:abab6c63-91d1-4f36-9be9-ae3b1a43742e nodeName:}" failed. No retries permitted until 2025-12-03 13:10:03.798318899 +0000 UTC m=+49.779239352 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs") pod "network-metrics-daemon-mbbpr" (UID: "abab6c63-91d1-4f36-9be9-ae3b1a43742e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.817721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.817778 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.817792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.817818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.817832 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.920778 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.920831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.920843 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.920889 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.920911 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.937831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.937900 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.937920 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.937951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.937976 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: E1203 13:09:55.954007 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.958064 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.958110 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.958124 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.958145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.958160 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: E1203 13:09:55.974692 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.979407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.979450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.979461 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.979483 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.979499 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:55 crc kubenswrapper[4690]: E1203 13:09:55.993520 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.997604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.997665 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.997691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.997739 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:55 crc kubenswrapper[4690]: I1203 13:09:55.997763 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:55Z","lastTransitionTime":"2025-12-03T13:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.314743 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.314774 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:56 crc kubenswrapper[4690]: E1203 13:09:56.315027 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:56 crc kubenswrapper[4690]: E1203 13:09:56.315218 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:56 crc kubenswrapper[4690]: E1203 13:09:56.704645 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.708857 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.708907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.708919 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.708939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.708950 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:56Z","lastTransitionTime":"2025-12-03T13:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:56 crc kubenswrapper[4690]: E1203 13:09:56.724419 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:56 crc kubenswrapper[4690]: E1203 13:09:56.724599 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.726630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.726667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.726682 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.726703 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.726717 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:56Z","lastTransitionTime":"2025-12-03T13:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.829588 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.829642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.829654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.829671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.829682 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:56Z","lastTransitionTime":"2025-12-03T13:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.932681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.932725 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.932735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.932755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:56 crc kubenswrapper[4690]: I1203 13:09:56.932765 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:56Z","lastTransitionTime":"2025-12-03T13:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.035069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.035108 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.035117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.035143 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.035155 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.141224 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.141957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.142014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.142052 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.142074 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.245274 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.245329 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.245341 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.245358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.245670 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.313703 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.313782 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:57 crc kubenswrapper[4690]: E1203 13:09:57.313892 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:09:57 crc kubenswrapper[4690]: E1203 13:09:57.314009 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.348463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.348494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.348511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.348529 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.348541 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.451031 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.451088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.451111 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.451142 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.451155 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.553499 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.553574 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.553591 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.553612 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.553627 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.655300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.655343 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.655358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.655378 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.655391 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.758034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.758223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.758256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.758334 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.758357 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.861682 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.861734 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.861747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.861769 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.861778 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.964034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.964075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.964086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.964102 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:57 crc kubenswrapper[4690]: I1203 13:09:57.964112 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:57Z","lastTransitionTime":"2025-12-03T13:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.065673 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.065718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.065727 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.065743 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.065754 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.168207 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.168254 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.168265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.168288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.168298 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.270897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.271126 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.271168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.271193 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.271213 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.314578 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.314676 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:09:58 crc kubenswrapper[4690]: E1203 13:09:58.314755 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:09:58 crc kubenswrapper[4690]: E1203 13:09:58.315076 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.315527 4690 scope.go:117] "RemoveContainer" containerID="e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.331824 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.348937 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.370320 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.374073 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.374277 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.374410 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.374508 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.374616 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.386747 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.399303 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.410388 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.423272 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.437897 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.456102 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.470249 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.477235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.477306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.477320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.477338 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.477350 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.495522 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.509607 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.523043 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.539235 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.560156 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"message\\\":\\\"d\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 13:09:47.297861 6116 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:09:47.297958 6116 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-webhook per-node LB for network=default: []services.LB{}\\\\nF1203 13:09:47.297967 6116 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.571747 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.580472 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.580525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.580546 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.580572 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.580590 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.582703 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.683360 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.683403 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.683414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.683432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.683443 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.787056 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.787115 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.787130 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.787155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.787174 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.890116 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.890175 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.890189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.890210 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.890223 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.993711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.993755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.993764 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.993782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:58 crc kubenswrapper[4690]: I1203 13:09:58.993792 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:58Z","lastTransitionTime":"2025-12-03T13:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.096980 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.097024 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.097035 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.097051 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.097063 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.200081 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.200126 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.200137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.200154 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.200164 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.303004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.303053 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.303065 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.303081 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.303093 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.314598 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.314673 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:09:59 crc kubenswrapper[4690]: E1203 13:09:59.314747 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:09:59 crc kubenswrapper[4690]: E1203 13:09:59.314896 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.405656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.405718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.405731 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.405750 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.405767 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.508286 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.508340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.508351 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.508376 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.508390 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.610706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.610756 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.610767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.610782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.610793 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.663832 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/1.log" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.666601 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.666977 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.680558 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.694044 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.708586 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.712579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.712691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.712757 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.712818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.712887 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.720532 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.740733 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.756464 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.769851 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.783412 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.798805 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.812394 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.815528 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.815584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.815603 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.815625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.815639 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.828029 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.848650 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"message\\\":\\\"d\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 13:09:47.297861 6116 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:09:47.297958 6116 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-webhook per-node LB for network=default: []services.LB{}\\\\nF1203 13:09:47.297967 6116 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.859945 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.880658 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.895531 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.908213 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.918800 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.918837 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.918847 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.918884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.918843 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:09:59 crc kubenswrapper[4690]: I1203 13:09:59.918903 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:09:59Z","lastTransitionTime":"2025-12-03T13:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.021759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.021794 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.021803 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.021818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.021829 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.124579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.124641 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.124655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.124680 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.124696 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.227329 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.227363 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.227382 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.227399 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.227409 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.314129 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.314129 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:00 crc kubenswrapper[4690]: E1203 13:10:00.314298 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:00 crc kubenswrapper[4690]: E1203 13:10:00.314324 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.330547 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.330602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.330616 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.330636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.330648 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.433584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.433650 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.433663 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.433690 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.433706 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.536735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.536786 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.536796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.536813 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.536829 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.639927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.640000 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.640016 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.640039 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.640054 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.672700 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/2.log" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.674068 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/1.log" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.677391 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d" exitCode=1 Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.677477 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.677541 4690 scope.go:117] "RemoveContainer" containerID="e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.679799 4690 scope.go:117] "RemoveContainer" containerID="f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d" Dec 03 13:10:00 crc kubenswrapper[4690]: E1203 13:10:00.680294 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.695593 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.713110 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.733043 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.743723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.743751 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.743778 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.743792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.743802 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.749155 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.764580 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.796440 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.812607 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.826193 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.840277 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.848967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.849037 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.849057 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.849085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.849107 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.866468 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4faa8ffcf37db9319abfb284fb91368145e8c6e9b8180c9389c248f2a3a4d06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"message\\\":\\\"d\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.254\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 13:09:47.297861 6116 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:09:47.297958 6116 services_controller.go:452] Built service openshift-machine-api/machine-api-operator-webhook per-node LB for network=default: []services.LB{}\\\\nF1203 13:09:47.297967 6116 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.885405 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.902493 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.917132 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.942672 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.952331 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.952401 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.952429 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.952467 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.952492 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:00Z","lastTransitionTime":"2025-12-03T13:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.960298 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.978906 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:00 crc kubenswrapper[4690]: I1203 13:10:00.996661 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.056139 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.056208 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.056233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.056265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.056290 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.160109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.160156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.160168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.160185 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.160197 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.263929 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.263966 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.263975 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.263992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.264003 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.314624 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.314700 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:01 crc kubenswrapper[4690]: E1203 13:10:01.314822 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:01 crc kubenswrapper[4690]: E1203 13:10:01.315036 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.368114 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.368183 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.368202 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.368230 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.368251 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.471240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.471309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.471321 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.471340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.471353 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.574102 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.574141 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.574152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.574168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.574179 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.676720 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.676833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.676844 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.676858 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.676884 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.681885 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/2.log" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.686190 4690 scope.go:117] "RemoveContainer" containerID="f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d" Dec 03 13:10:01 crc kubenswrapper[4690]: E1203 13:10:01.686428 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.708434 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.733454 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.748351 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.763936 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.779627 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.779688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.779703 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.779726 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.779744 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.782348 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.796544 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.811276 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.827263 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.843558 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.859585 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.882559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.882605 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.882666 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.882688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.882727 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.882578 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.895180 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.921389 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.936321 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.948221 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.962118 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.973770 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.985940 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.985991 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.986001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.986022 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:01 crc kubenswrapper[4690]: I1203 13:10:01.986037 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:01Z","lastTransitionTime":"2025-12-03T13:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.092140 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.092186 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.092196 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.092213 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.092224 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.195478 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.195515 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.195527 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.195544 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.195554 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.298002 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.298043 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.298056 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.298073 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.298085 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.314775 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.314802 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:02 crc kubenswrapper[4690]: E1203 13:10:02.315196 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:02 crc kubenswrapper[4690]: E1203 13:10:02.315331 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.400659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.400741 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.400761 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.400796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.400819 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.503856 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.503975 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.503994 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.504022 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.504043 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.607287 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.607326 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.607335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.607350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.607361 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.710395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.710439 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.710451 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.710466 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.710478 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.812702 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.812740 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.812753 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.812772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.812784 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.915656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.915716 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.915728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.915749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:02 crc kubenswrapper[4690]: I1203 13:10:02.915771 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:02Z","lastTransitionTime":"2025-12-03T13:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.019277 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.019351 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.019367 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.019395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.019417 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.123108 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.123185 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.123198 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.123220 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.123237 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.225791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.225886 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.225901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.225926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.225944 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.314470 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.314487 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:03 crc kubenswrapper[4690]: E1203 13:10:03.314698 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:03 crc kubenswrapper[4690]: E1203 13:10:03.314812 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.329004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.329355 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.329434 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.329504 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.329565 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.433581 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.433633 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.433650 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.433666 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.433677 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.536533 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.536584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.536597 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.536613 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.536624 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.639961 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.640005 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.640014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.640030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.640041 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.742667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.742720 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.742731 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.742775 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.742792 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.845710 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.845758 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.845771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.845791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.845816 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.892542 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:03 crc kubenswrapper[4690]: E1203 13:10:03.892713 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:10:03 crc kubenswrapper[4690]: E1203 13:10:03.892782 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs podName:abab6c63-91d1-4f36-9be9-ae3b1a43742e nodeName:}" failed. No retries permitted until 2025-12-03 13:10:19.89276213 +0000 UTC m=+65.873682563 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs") pod "network-metrics-daemon-mbbpr" (UID: "abab6c63-91d1-4f36-9be9-ae3b1a43742e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.949314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.949364 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.949374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.949392 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:03 crc kubenswrapper[4690]: I1203 13:10:03.949403 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:03Z","lastTransitionTime":"2025-12-03T13:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.051467 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.051509 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.051519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.051536 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.051547 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.153832 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.153891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.153903 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.153918 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.153928 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.195924 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.196064 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:10:36.196043816 +0000 UTC m=+82.176964249 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.196102 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.196125 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.196240 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.196252 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.196295 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:10:36.196285502 +0000 UTC m=+82.177205935 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.196339 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:10:36.196331743 +0000 UTC m=+82.177252186 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.256337 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.256374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.256385 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.256401 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.256415 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.297606 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.297654 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.297774 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.297792 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.297804 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.297850 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.297903 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.297915 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.297859 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:10:36.297843951 +0000 UTC m=+82.278764384 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.297981 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:10:36.297961774 +0000 UTC m=+82.278882307 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.314444 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.314574 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.314913 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:04 crc kubenswrapper[4690]: E1203 13:10:04.314991 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.329127 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.341286 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.355210 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.359317 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.359386 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.359399 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.359423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.359437 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.375137 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.388087 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.400048 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.420303 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.433540 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.457580 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.462161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.462195 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.462204 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.462221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.462231 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.475616 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.490495 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.508328 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.526777 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.543325 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.557312 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.565679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.565718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.565729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.565746 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.565758 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.572117 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.589532 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.669044 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.669110 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.669121 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.669144 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.669157 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.772207 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.772249 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.772259 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.772275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.772287 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.874335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.874370 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.874380 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.874395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.874405 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.977503 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.977572 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.977586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.977604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:04 crc kubenswrapper[4690]: I1203 13:10:04.977616 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:04Z","lastTransitionTime":"2025-12-03T13:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.081379 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.081440 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.081455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.081474 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.081485 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.184672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.184718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.184731 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.184747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.184758 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.287415 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.287453 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.287462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.287477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.287487 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.314325 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.314397 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:05 crc kubenswrapper[4690]: E1203 13:10:05.314470 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:05 crc kubenswrapper[4690]: E1203 13:10:05.314631 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.390212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.390244 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.390254 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.390269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.390278 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.492972 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.493011 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.493022 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.493039 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.493052 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.594968 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.595003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.595014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.595029 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.595038 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.697766 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.697833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.697844 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.697883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.697898 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.801004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.801089 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.801109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.801137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.801158 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.904203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.904245 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.904256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.904272 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:05 crc kubenswrapper[4690]: I1203 13:10:05.904280 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:05Z","lastTransitionTime":"2025-12-03T13:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.006707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.006746 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.006755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.006770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.006779 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.109654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.109703 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.109715 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.109740 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.109754 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.213160 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.213597 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.213626 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.213654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.213674 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.314035 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.314250 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:06 crc kubenswrapper[4690]: E1203 13:10:06.314968 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:06 crc kubenswrapper[4690]: E1203 13:10:06.315317 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.317008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.317058 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.317069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.317088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.317102 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.421770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.422257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.422337 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.422417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.422476 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.526321 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.526411 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.526435 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.526467 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.526489 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.629189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.629256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.629272 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.629293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.629305 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.731937 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.732228 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.732291 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.732412 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.732474 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.835738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.835917 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.835931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.835948 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.835961 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.837092 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.837278 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.837418 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.837552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.837672 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: E1203 13:10:06.858719 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.863602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.863645 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.863654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.863670 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.863681 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: E1203 13:10:06.879025 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.884514 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.884571 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.884590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.884616 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.884633 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: E1203 13:10:06.898299 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.904326 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.904362 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.904374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.904396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.904408 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: E1203 13:10:06.918632 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.923663 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.923720 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.923738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.923767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.923779 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:06 crc kubenswrapper[4690]: E1203 13:10:06.945587 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:06 crc kubenswrapper[4690]: E1203 13:10:06.945831 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.948211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.948253 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.948264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.948283 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:06 crc kubenswrapper[4690]: I1203 13:10:06.948294 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:06Z","lastTransitionTime":"2025-12-03T13:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.052321 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.052370 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.052380 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.052398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.052449 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.158200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.158254 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.158269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.158290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.158304 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.261697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.261749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.261759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.261777 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.261787 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.313836 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:07 crc kubenswrapper[4690]: E1203 13:10:07.314129 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.314285 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:07 crc kubenswrapper[4690]: E1203 13:10:07.314525 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.365525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.365587 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.365600 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.365618 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.365630 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.468419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.468546 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.468566 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.468589 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.468604 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.571670 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.571736 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.571750 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.572075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.572103 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.674710 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.674806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.674819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.674836 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.674848 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.719565 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.732577 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.743603 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.757360 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.778359 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.778416 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.778428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.778450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.778464 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.786633 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.804785 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.829024 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.844566 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.858710 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.871530 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.882801 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.882850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.882861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.882900 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.882912 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.886962 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.900018 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.915246 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.927570 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.941024 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.958291 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.975673 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.985738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.985804 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.985823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.985850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.985912 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:07Z","lastTransitionTime":"2025-12-03T13:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:07 crc kubenswrapper[4690]: I1203 13:10:07.988591 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.000801 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.088178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.088475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.088553 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.088639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.088697 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.190959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.191004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.191014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.191032 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.191041 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.293264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.293665 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.293861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.294126 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.294332 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.314750 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:08 crc kubenswrapper[4690]: E1203 13:10:08.314925 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.314767 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:08 crc kubenswrapper[4690]: E1203 13:10:08.315098 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.396982 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.397026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.397037 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.397055 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.397067 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.500250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.500301 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.500312 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.500330 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.500341 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.602810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.602856 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.602889 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.602910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.602922 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.705441 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.705517 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.705553 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.705583 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.705603 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.809201 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.809265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.809283 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.809311 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.809332 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.911728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.911777 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.911785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.911802 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:08 crc kubenswrapper[4690]: I1203 13:10:08.911812 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:08Z","lastTransitionTime":"2025-12-03T13:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.014453 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.014499 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.014508 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.014526 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.014535 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.116738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.116783 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.116793 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.116812 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.116822 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.219647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.219699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.219709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.219726 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.219740 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.314858 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.314916 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:09 crc kubenswrapper[4690]: E1203 13:10:09.315110 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:09 crc kubenswrapper[4690]: E1203 13:10:09.315371 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.323014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.323061 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.323075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.323099 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.323111 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.425601 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.425656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.425670 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.425692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.425707 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.528721 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.528780 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.528790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.528813 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.528827 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.631419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.631463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.631471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.631486 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.631496 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.734665 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.734716 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.734728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.734748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.734761 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.837723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.837785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.837797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.837815 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.837827 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.940935 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.940997 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.941008 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.941036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:09 crc kubenswrapper[4690]: I1203 13:10:09.941058 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:09Z","lastTransitionTime":"2025-12-03T13:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.044407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.044454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.044467 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.044487 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.044503 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.147625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.147737 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.147767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.147795 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.147813 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.250380 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.250450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.250474 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.250505 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.250532 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.315198 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.315346 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:10 crc kubenswrapper[4690]: E1203 13:10:10.315426 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:10 crc kubenswrapper[4690]: E1203 13:10:10.315548 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.353359 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.353409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.353421 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.353442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.353457 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.455489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.455554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.455567 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.455583 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.455594 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.558342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.558385 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.558397 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.558412 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.558421 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.662487 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.662570 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.662592 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.662638 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.662653 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.765636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.765675 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.765684 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.765700 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.765720 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.869000 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.869048 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.869059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.869079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.869095 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.972363 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.972419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.972428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.972445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:10 crc kubenswrapper[4690]: I1203 13:10:10.972455 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:10Z","lastTransitionTime":"2025-12-03T13:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.077079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.077134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.077146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.077177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.077193 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.181349 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.181429 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.181452 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.181486 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.181513 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.285189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.285249 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.285261 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.285281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.285295 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.313773 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:11 crc kubenswrapper[4690]: E1203 13:10:11.313987 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.314273 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:11 crc kubenswrapper[4690]: E1203 13:10:11.314348 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.387985 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.388032 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.388043 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.388060 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.388073 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.492095 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.492162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.492172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.492191 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.492204 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.595475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.595542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.595555 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.595584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.595604 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.699046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.699101 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.699112 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.699134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.699147 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.802575 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.802625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.802636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.802654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.802668 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.905886 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.905935 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.905946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.905964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:11 crc kubenswrapper[4690]: I1203 13:10:11.905977 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:11Z","lastTransitionTime":"2025-12-03T13:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.008138 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.008200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.008214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.008233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.008245 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.111535 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.111632 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.111656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.111691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.111715 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.214810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.214967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.214983 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.215004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.215017 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.313756 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.313795 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:12 crc kubenswrapper[4690]: E1203 13:10:12.313923 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:12 crc kubenswrapper[4690]: E1203 13:10:12.314000 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.317684 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.317720 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.317729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.317763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.317773 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.421578 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.421659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.421673 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.421691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.421735 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.525009 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.525056 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.525069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.525084 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.525096 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.628480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.628553 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.628572 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.628593 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.628609 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.732100 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.732167 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.732186 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.732232 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.732245 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.835131 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.835197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.835215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.835240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.835261 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.938336 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.938399 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.938412 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.938438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:12 crc kubenswrapper[4690]: I1203 13:10:12.938456 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:12Z","lastTransitionTime":"2025-12-03T13:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.041654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.041697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.041706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.041726 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.041737 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.145142 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.145208 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.145222 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.145242 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.145276 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.247846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.247910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.247919 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.247932 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.247941 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.313960 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.313998 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:13 crc kubenswrapper[4690]: E1203 13:10:13.314112 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:13 crc kubenswrapper[4690]: E1203 13:10:13.314299 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.314976 4690 scope.go:117] "RemoveContainer" containerID="f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d" Dec 03 13:10:13 crc kubenswrapper[4690]: E1203 13:10:13.315171 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.355961 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.356014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.356264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.356290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.356304 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.460695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.460757 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.460778 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.460808 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.460832 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.564134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.564210 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.564237 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.564270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.564291 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.667569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.667645 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.667662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.667685 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.667704 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.770206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.770265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.770276 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.770313 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.770325 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.872598 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.872657 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.872668 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.872700 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.872712 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.975699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.975743 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.975752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.975769 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:13 crc kubenswrapper[4690]: I1203 13:10:13.975780 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:13Z","lastTransitionTime":"2025-12-03T13:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.077625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.077672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.077684 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.077700 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.077711 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.180424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.180483 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.180494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.180513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.180526 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.282812 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.282882 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.282899 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.282918 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.282933 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.313988 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:14 crc kubenswrapper[4690]: E1203 13:10:14.314111 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.314274 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:14 crc kubenswrapper[4690]: E1203 13:10:14.314387 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.329170 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.340884 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.359798 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.370697 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.386268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.386320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.386342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.386370 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.386387 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.387805 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.402789 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.414998 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.426255 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.439810 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.453601 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.468891 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.479133 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.489283 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.489339 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.489353 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.489376 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.489389 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.492613 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.509525 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.524431 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.537494 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.554468 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.573671 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.592383 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.592414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.592423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.592436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.592445 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.695056 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.695104 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.695114 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.695129 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.695139 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.799178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.799237 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.799248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.799268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.799280 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.901694 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.901749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.901759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.901775 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:14 crc kubenswrapper[4690]: I1203 13:10:14.901788 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:14Z","lastTransitionTime":"2025-12-03T13:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.007112 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.007182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.007210 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.007233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.007247 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.109207 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.109248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.109260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.109276 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.109298 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.212255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.212304 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.212317 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.212332 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.212341 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.313713 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.313752 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:15 crc kubenswrapper[4690]: E1203 13:10:15.313896 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:15 crc kubenswrapper[4690]: E1203 13:10:15.314004 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.316424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.316455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.316473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.316492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.316506 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.419662 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.419709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.419719 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.419738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.419749 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.522178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.522234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.522245 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.522269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.522286 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.625145 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.625181 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.625189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.625203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.625212 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.728049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.728109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.728123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.728142 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.728158 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.830997 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.831050 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.831061 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.831079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.831089 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.933504 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.933547 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.933556 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.933571 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:15 crc kubenswrapper[4690]: I1203 13:10:15.933580 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:15Z","lastTransitionTime":"2025-12-03T13:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.036972 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.037026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.037038 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.037057 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.037071 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.139314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.139360 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.139370 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.139388 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.139401 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.242580 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.242617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.242630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.242645 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.242664 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.314548 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.314613 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:16 crc kubenswrapper[4690]: E1203 13:10:16.314708 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:16 crc kubenswrapper[4690]: E1203 13:10:16.314802 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.345564 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.345634 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.345650 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.345678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.345696 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.449172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.449261 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.449284 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.449321 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.449355 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.552227 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.552288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.552301 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.552317 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.552329 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.654592 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.654630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.654647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.654661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.654669 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.758282 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.758345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.758359 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.758377 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.758392 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.862522 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.862625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.862646 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.862720 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.862741 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.965697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.965735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.965749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.965763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.965773 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.969055 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.969105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.969117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.969133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.969144 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:16 crc kubenswrapper[4690]: E1203 13:10:16.982781 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.988255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.988303 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.988321 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.988337 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:16 crc kubenswrapper[4690]: I1203 13:10:16.988346 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:16Z","lastTransitionTime":"2025-12-03T13:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: E1203 13:10:17.010047 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.015851 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.015905 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.015919 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.015936 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.015949 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: E1203 13:10:17.030925 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.035110 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.035158 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.035166 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.035187 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.035199 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: E1203 13:10:17.051350 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.056081 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.056120 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.056135 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.056156 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.056170 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: E1203 13:10:17.071362 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:17 crc kubenswrapper[4690]: E1203 13:10:17.071525 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.073262 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.073285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.073295 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.073313 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.073324 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.175464 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.175516 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.175530 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.175552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.175567 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.279335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.279420 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.279443 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.279477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.279501 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.314608 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.314696 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:17 crc kubenswrapper[4690]: E1203 13:10:17.314845 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:17 crc kubenswrapper[4690]: E1203 13:10:17.315032 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.382298 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.382351 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.382368 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.382386 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.382399 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.484810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.484852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.484878 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.484894 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.484904 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.587542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.587581 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.587594 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.587611 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.587622 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.689825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.689861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.689891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.689907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.689920 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.792286 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.792331 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.792344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.792360 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.792370 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.894950 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.894989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.894998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.895015 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.895024 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.999072 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.999123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.999133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.999151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:17 crc kubenswrapper[4690]: I1203 13:10:17.999167 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:17Z","lastTransitionTime":"2025-12-03T13:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.101350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.101390 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.101401 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.101417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.101428 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.204032 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.204067 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.204076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.204090 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.204100 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.306297 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.306342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.306358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.306378 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.306390 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.314594 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.314598 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:18 crc kubenswrapper[4690]: E1203 13:10:18.314760 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:18 crc kubenswrapper[4690]: E1203 13:10:18.314694 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.409013 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.409051 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.409059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.409076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.409089 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.512206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.512264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.512273 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.512292 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.512302 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.615211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.615247 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.615256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.615272 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.615283 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.717605 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.718128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.718263 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.718409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.718546 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.821921 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.821982 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.822003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.822026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.822043 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.924428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.924481 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.924493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.924512 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:18 crc kubenswrapper[4690]: I1203 13:10:18.924525 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:18Z","lastTransitionTime":"2025-12-03T13:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.027069 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.027129 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.027140 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.027158 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.027168 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.129709 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.129747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.129759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.129776 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.129786 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.232929 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.233004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.233022 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.233052 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.233071 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.314534 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.314639 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:19 crc kubenswrapper[4690]: E1203 13:10:19.314692 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:19 crc kubenswrapper[4690]: E1203 13:10:19.314809 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.335784 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.335845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.335880 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.335900 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.335912 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.438161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.438198 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.438208 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.438221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.438230 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.540748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.540799 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.540809 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.540827 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.541215 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.644603 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.644663 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.644677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.644697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.644710 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.747439 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.747539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.747549 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.747566 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.747576 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.851023 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.851060 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.851071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.851088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.851103 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.953841 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.953919 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.953935 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.954311 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.954356 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:19Z","lastTransitionTime":"2025-12-03T13:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:19 crc kubenswrapper[4690]: I1203 13:10:19.975677 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:19 crc kubenswrapper[4690]: E1203 13:10:19.975849 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:10:19 crc kubenswrapper[4690]: E1203 13:10:19.975944 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs podName:abab6c63-91d1-4f36-9be9-ae3b1a43742e nodeName:}" failed. No retries permitted until 2025-12-03 13:10:51.975927367 +0000 UTC m=+97.956847800 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs") pod "network-metrics-daemon-mbbpr" (UID: "abab6c63-91d1-4f36-9be9-ae3b1a43742e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.056446 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.056489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.056501 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.056516 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.056525 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.159236 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.159268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.159277 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.159291 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.159301 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.261938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.261984 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.261998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.262016 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.262029 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.314076 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.314162 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:20 crc kubenswrapper[4690]: E1203 13:10:20.314198 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:20 crc kubenswrapper[4690]: E1203 13:10:20.314236 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.364814 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.364855 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.364913 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.364931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.364982 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.467256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.467292 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.467305 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.467321 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.467333 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.570203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.570249 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.570259 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.570275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.570286 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.673081 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.673119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.673128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.673143 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.673152 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.746418 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/0.log" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.746475 4690 generic.go:334] "Generic (PLEG): container finished" podID="bf580211-e03a-444e-98b8-c12f99091b80" containerID="7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a" exitCode=1 Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.746511 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tf2cb" event={"ID":"bf580211-e03a-444e-98b8-c12f99091b80","Type":"ContainerDied","Data":"7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.746997 4690 scope.go:117] "RemoveContainer" containerID="7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.759084 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.771199 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.775481 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.775528 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.775539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.775558 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.775572 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.783268 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.795284 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.809749 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.825582 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.837634 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.848155 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.866005 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.880333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.880625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.880924 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.880969 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.881014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.881130 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.893776 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.904121 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.921174 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.933426 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.952762 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.966241 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.979699 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.984117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.984163 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.984177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.984197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.984213 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:20Z","lastTransitionTime":"2025-12-03T13:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:20 crc kubenswrapper[4690]: I1203 13:10:20.995237 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.087196 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.087720 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.087736 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.087759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.087775 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.190696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.191034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.191112 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.191193 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.191306 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.294496 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.294547 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.294557 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.294573 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.294584 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.313955 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.314057 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:21 crc kubenswrapper[4690]: E1203 13:10:21.314115 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:21 crc kubenswrapper[4690]: E1203 13:10:21.314244 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.398097 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.398146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.398159 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.398179 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.398193 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.501402 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.501446 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.501455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.501474 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.501486 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.603595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.603648 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.603658 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.603672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.603697 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.707943 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.708004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.708018 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.708035 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.708048 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.751247 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/0.log" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.751316 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tf2cb" event={"ID":"bf580211-e03a-444e-98b8-c12f99091b80","Type":"ContainerStarted","Data":"f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.771452 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.792944 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.811166 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.811221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.811234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.811255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.811267 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.811272 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.831455 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.844494 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.860288 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.878433 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.892613 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.908505 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.913254 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.913319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.913335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.913357 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.913370 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:21Z","lastTransitionTime":"2025-12-03T13:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.932529 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.946242 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.962159 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.976277 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:21 crc kubenswrapper[4690]: I1203 13:10:21.997919 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.014695 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.016422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.016503 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.016516 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.016544 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.016558 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.029383 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.043284 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.055789 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.119373 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.119414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.119424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.119440 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.119450 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.222569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.222637 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.222650 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.222669 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.222682 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.314096 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.314128 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:22 crc kubenswrapper[4690]: E1203 13:10:22.314225 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:22 crc kubenswrapper[4690]: E1203 13:10:22.314431 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.324747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.324790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.324802 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.324817 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.324829 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.427384 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.427426 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.427436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.427452 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.427461 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.530567 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.530614 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.530626 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.530647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.530662 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.633206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.633257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.633269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.633290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.633321 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.736175 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.736214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.736223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.736235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.736244 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.837939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.837976 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.837985 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.837999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.838008 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.940251 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.940284 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.940294 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.940307 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:22 crc kubenswrapper[4690]: I1203 13:10:22.940317 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:22Z","lastTransitionTime":"2025-12-03T13:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.042627 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.042659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.042668 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.042681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.042690 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.145580 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.145655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.145671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.145688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.145700 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.248818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.248879 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.248925 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.248944 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.248957 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.314185 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.314221 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:23 crc kubenswrapper[4690]: E1203 13:10:23.314349 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:23 crc kubenswrapper[4690]: E1203 13:10:23.314446 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.351172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.351214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.351224 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.351244 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.351255 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.454109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.454155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.454171 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.454189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.454200 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.556482 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.556536 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.556548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.556568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.556580 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.659026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.659118 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.659136 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.659223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.659263 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.761355 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.761419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.761436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.761460 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.761477 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.864020 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.864066 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.864079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.864097 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.864107 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.966948 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.966992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.967003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.967019 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:23 crc kubenswrapper[4690]: I1203 13:10:23.967029 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:23Z","lastTransitionTime":"2025-12-03T13:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.070591 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.070646 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.070656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.070675 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.070687 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.173402 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.173450 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.173460 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.173479 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.173496 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.276236 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.276291 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.276305 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.276322 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.276335 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.314391 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.314417 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:24 crc kubenswrapper[4690]: E1203 13:10:24.314578 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:24 crc kubenswrapper[4690]: E1203 13:10:24.314812 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.330367 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.350253 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.364122 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.378084 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.379519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.379572 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.379590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.379616 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.379635 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.396935 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.409761 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.424542 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.446205 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.458030 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.479887 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.483909 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.483968 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.483987 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.484010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.484022 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.496462 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.509522 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.525164 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.537159 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.549657 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.561504 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.571808 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.587696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.587789 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.587830 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.587856 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.587899 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.588032 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.691081 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.691129 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.691138 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.691155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.691167 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.794311 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.794791 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.795147 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.795417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.795677 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.899226 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.899720 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.899921 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.900086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:24 crc kubenswrapper[4690]: I1203 13:10:24.900225 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:24Z","lastTransitionTime":"2025-12-03T13:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.002732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.002774 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.002787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.002814 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.002835 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.106585 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.107081 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.107295 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.107514 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.107622 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.210295 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.210844 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.211057 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.211240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.211389 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.313753 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.313813 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:25 crc kubenswrapper[4690]: E1203 13:10:25.314180 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:25 crc kubenswrapper[4690]: E1203 13:10:25.314294 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.314649 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.314756 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.314842 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.314943 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.315037 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.418803 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.419197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.419329 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.419434 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.419526 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.522913 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.523680 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.523751 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.523838 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.523935 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.626258 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.626300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.626314 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.626332 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.626343 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.728628 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.728676 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.728688 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.728707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.728721 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.831944 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.832286 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.832377 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.832473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.832561 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.935260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.935558 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.935635 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.935712 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:25 crc kubenswrapper[4690]: I1203 13:10:25.935828 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:25Z","lastTransitionTime":"2025-12-03T13:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.037891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.037928 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.037938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.037976 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.037992 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.141191 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.141528 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.141663 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.141759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.141851 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.243819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.243880 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.243895 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.243912 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.243922 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.314006 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.314054 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:26 crc kubenswrapper[4690]: E1203 13:10:26.314581 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:26 crc kubenswrapper[4690]: E1203 13:10:26.314697 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.315499 4690 scope.go:117] "RemoveContainer" containerID="f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.325740 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.347087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.347400 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.347477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.347548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.347616 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.450409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.450448 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.450475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.450495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.450506 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.552942 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.553578 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.553960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.554087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.554208 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.656579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.656903 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.656994 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.657089 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.657223 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.760405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.760445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.760455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.760472 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.760485 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.863114 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.863179 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.863193 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.863222 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.863239 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.965706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.965755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.965764 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.965782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:26 crc kubenswrapper[4690]: I1203 13:10:26.965793 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:26Z","lastTransitionTime":"2025-12-03T13:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.067904 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.067947 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.067991 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.068010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.068020 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.121711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.122047 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.122168 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.122262 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.122352 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: E1203 13:10:27.135400 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.140061 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.140133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.140151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.140178 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.140196 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: E1203 13:10:27.153950 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.158649 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.158700 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.158711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.158772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.158787 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: E1203 13:10:27.170775 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.175152 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.175194 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.175211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.175229 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.175240 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: E1203 13:10:27.188392 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.192304 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.192358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.192369 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.192391 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.192406 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: E1203 13:10:27.205622 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: E1203 13:10:27.205763 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.207421 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.207467 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.207478 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.207493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.207505 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.310467 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.310523 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.310533 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.310555 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.310566 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.315521 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:27 crc kubenswrapper[4690]: E1203 13:10:27.315822 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.316006 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:27 crc kubenswrapper[4690]: E1203 13:10:27.316664 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.413183 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.413234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.413249 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.413267 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.413280 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.516355 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.517118 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.517164 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.517190 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.517204 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.621549 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.621598 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.621611 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.621633 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.621648 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.724233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.724269 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.724279 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.724295 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.724306 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.772189 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/2.log" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.775446 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.776058 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.794605 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.810463 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.823999 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.826937 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.826986 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.826997 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.827015 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.827027 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.840617 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.854322 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.868197 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afdda9eb-0b52-4293-a70f-7ee5a507df8b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9a2357cbe74fa94cec46032822a70d9124c382e568d434ec7dfc64167aa8aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.881650 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.899306 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.909929 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.921654 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.929527 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.929559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.929570 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.929584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.929594 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:27Z","lastTransitionTime":"2025-12-03T13:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.941176 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.959854 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:27 crc kubenswrapper[4690]: I1203 13:10:27.978732 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.002310 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.020601 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.032114 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.032155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.032172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.032188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.032196 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.033932 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.045455 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.067773 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.079178 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.134247 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.134297 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.134306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.134319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.134329 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.236903 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.236952 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.236973 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.236990 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.237003 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.314376 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.314444 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:28 crc kubenswrapper[4690]: E1203 13:10:28.314538 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:28 crc kubenswrapper[4690]: E1203 13:10:28.314800 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.340042 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.340078 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.340088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.340104 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.340114 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.442926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.442974 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.442985 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.443003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.443012 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.545308 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.545347 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.545357 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.545373 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.545384 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.647418 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.647452 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.647460 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.647476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.647486 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.750283 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.750332 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.750342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.750360 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.750371 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.781324 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/3.log" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.781935 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/2.log" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.784203 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" exitCode=1 Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.784243 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.784290 4690 scope.go:117] "RemoveContainer" containerID="f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.784838 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:10:28 crc kubenswrapper[4690]: E1203 13:10:28.785029 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.803017 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.815593 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.826702 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.842640 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.853244 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.853290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.853303 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.853324 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.853338 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.857449 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.871156 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afdda9eb-0b52-4293-a70f-7ee5a507df8b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9a2357cbe74fa94cec46032822a70d9124c382e568d434ec7dfc64167aa8aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.889641 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.906677 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.918820 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.933604 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.952421 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.956392 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.956445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.956455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.956478 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.956490 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:28Z","lastTransitionTime":"2025-12-03T13:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.966075 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.977298 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:28 crc kubenswrapper[4690]: I1203 13:10:28.996356 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:28Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.011524 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.024739 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.035459 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.056031 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9009a1466f0196c1a9e00c619e02580e4b1ac1500a46b481c5541b229e5c43d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:00Z\\\",\\\"message\\\":\\\"7 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:09:59.613625 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 13:09:59.613632 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:09:59.613641 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:09:59.613664 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.613975 6308 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614186 6308 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 13:09:59.614317 6308 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:09:59.614766 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:09:59.614786 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:09:59.614812 6308 factory.go:656] Stopping watch factory\\\\nI1203 13:09:59.614830 6308 ovnkube.go:599] Stopped ovnkube\\\\nI1203 13:09:59.614926 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:28Z\\\",\\\"message\\\":\\\":10:27.998491 6669 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:10:27.996511 6669 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:10:27.998477 6669 services_controller.go:453] Built service openshift-network-console/networking-console-plugin template LB for network=default: []services.LB{}\\\\nF1203 13:10:27.998530 6669 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start nod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:10:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.058640 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.058678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.058689 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.058705 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.058725 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.069207 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.160429 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.160468 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.160477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.160492 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.160503 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.263092 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.263135 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.263144 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.263161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.263172 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.314324 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.314399 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:29 crc kubenswrapper[4690]: E1203 13:10:29.314478 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:29 crc kubenswrapper[4690]: E1203 13:10:29.314562 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.366268 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.366300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.366310 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.366324 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.366334 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.470280 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.470525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.470539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.470568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.470584 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.575399 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.575485 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.575511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.575551 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.575576 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.679172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.679236 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.679247 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.679267 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.679280 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.782030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.782116 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.782139 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.782169 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.782191 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.790087 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/3.log" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.794444 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:10:29 crc kubenswrapper[4690]: E1203 13:10:29.794675 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.807630 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.823819 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.838370 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.852697 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.868228 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.888321 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.892042 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.892139 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.892155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.892183 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.892208 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.906847 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.922717 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afdda9eb-0b52-4293-a70f-7ee5a507df8b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9a2357cbe74fa94cec46032822a70d9124c382e568d434ec7dfc64167aa8aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.937972 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.953057 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.971452 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.988805 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.995503 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.995558 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.995569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.995586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:29 crc kubenswrapper[4690]: I1203 13:10:29.995598 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:29Z","lastTransitionTime":"2025-12-03T13:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.003887 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.025333 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:28Z\\\",\\\"message\\\":\\\":10:27.998491 6669 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:10:27.996511 6669 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:10:27.998477 6669 services_controller.go:453] Built service openshift-network-console/networking-console-plugin template LB for network=default: []services.LB{}\\\\nF1203 13:10:27.998530 6669 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start nod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:10:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.040779 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.063047 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.081602 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.098971 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.099207 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.099267 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.099286 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.099310 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.099331 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.115095 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.202837 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.202944 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.202968 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.203001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.203021 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.306266 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.306352 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.306374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.306407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.306427 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.314488 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.314599 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:30 crc kubenswrapper[4690]: E1203 13:10:30.314652 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:30 crc kubenswrapper[4690]: E1203 13:10:30.314789 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.410131 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.410185 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.410199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.410218 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.410230 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.512643 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.512695 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.512710 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.512730 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.512743 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.615797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.615841 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.615859 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.615915 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.615929 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.718903 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.718963 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.718978 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.719003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.719016 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.821840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.821910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.821924 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.821943 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.821956 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.925149 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.925199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.925211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.925227 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:30 crc kubenswrapper[4690]: I1203 13:10:30.925238 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:30Z","lastTransitionTime":"2025-12-03T13:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.028248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.028301 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.028313 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.028333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.028348 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.131403 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.131454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.131469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.131490 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.131502 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.234610 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.234646 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.234654 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.234669 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.234678 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.314244 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.314323 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:31 crc kubenswrapper[4690]: E1203 13:10:31.314377 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:31 crc kubenswrapper[4690]: E1203 13:10:31.314526 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.336747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.336831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.336843 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.336884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.336898 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.439779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.439833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.439845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.439884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.439899 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.542795 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.542852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.542883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.542901 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.542913 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.645906 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.645964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.645974 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.645992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.646003 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.749262 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.749334 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.749345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.749366 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.749379 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.853153 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.853212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.853225 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.853252 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.853269 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.957854 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.957923 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.957935 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.957957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:31 crc kubenswrapper[4690]: I1203 13:10:31.957969 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:31Z","lastTransitionTime":"2025-12-03T13:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.061407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.061503 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.061517 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.061536 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.061548 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.164197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.164256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.164266 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.164285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.164298 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.267303 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.267343 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.267355 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.267374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.267385 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.314442 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.314516 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:32 crc kubenswrapper[4690]: E1203 13:10:32.314656 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:32 crc kubenswrapper[4690]: E1203 13:10:32.314822 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.370622 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.370667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.370682 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.370699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.370711 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.474214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.474288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.474309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.474343 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.474366 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.576570 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.576624 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.576639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.576659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.576673 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.679255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.679328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.679341 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.679358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.679370 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.782663 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.783593 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.783773 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.783960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.784092 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.887637 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.887713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.887731 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.887763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.887781 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.991346 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.991392 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.991405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.991422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:32 crc kubenswrapper[4690]: I1203 13:10:32.991435 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:32Z","lastTransitionTime":"2025-12-03T13:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.093970 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.094036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.094055 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.094080 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.094097 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.197884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.197945 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.197955 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.197979 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.197992 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.301575 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.302300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.302330 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.302368 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.302393 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.313960 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.314040 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:33 crc kubenswrapper[4690]: E1203 13:10:33.314168 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:33 crc kubenswrapper[4690]: E1203 13:10:33.314238 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.405308 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.405362 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.405373 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.405397 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.405410 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.507936 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.507980 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.507989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.508003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.512152 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.615534 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.615586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.615599 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.615615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.615624 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.717716 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.717910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.717921 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.717937 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.717947 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.820810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.820850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.820878 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.820897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.820909 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.935575 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.935627 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.935639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.935657 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:33 crc kubenswrapper[4690]: I1203 13:10:33.935670 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:33Z","lastTransitionTime":"2025-12-03T13:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.038365 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.038424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.038442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.038468 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.038485 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.142398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.142462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.142475 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.142494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.142506 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.244718 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.244793 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.244810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.245151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.245161 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.313995 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.314079 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:34 crc kubenswrapper[4690]: E1203 13:10:34.314145 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:34 crc kubenswrapper[4690]: E1203 13:10:34.314242 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.328463 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.347620 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.347668 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.347680 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.347699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.347719 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.348622 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:28Z\\\",\\\"message\\\":\\\":10:27.998491 6669 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:10:27.996511 6669 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:10:27.998477 6669 services_controller.go:453] Built service openshift-network-console/networking-console-plugin template LB for network=default: []services.LB{}\\\\nF1203 13:10:27.998530 6669 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start nod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:10:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.358489 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.376545 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.389088 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.400822 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.412472 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.425118 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.434005 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.443890 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.450004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.450248 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.450487 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.450689 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.450923 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.461789 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.471956 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.486886 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.499421 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.511013 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afdda9eb-0b52-4293-a70f-7ee5a507df8b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9a2357cbe74fa94cec46032822a70d9124c382e568d434ec7dfc64167aa8aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.523490 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.537260 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.552646 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.557707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.557752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.557762 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.557779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.557790 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.565090 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:34Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.660363 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.660417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.660434 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.660459 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.660477 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.763463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.763529 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.763546 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.763568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.763580 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.865908 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.865957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.865968 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.865984 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.865993 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.969271 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.969313 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.969325 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.969344 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:34 crc kubenswrapper[4690]: I1203 13:10:34.969357 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:34Z","lastTransitionTime":"2025-12-03T13:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.072263 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.072307 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.072320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.072338 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.072349 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.175234 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.175271 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.175282 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.175297 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.175310 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.277704 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.277749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.277761 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.277777 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.277788 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.314506 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:35 crc kubenswrapper[4690]: E1203 13:10:35.314639 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.314853 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:35 crc kubenswrapper[4690]: E1203 13:10:35.314932 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.380553 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.380600 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.380612 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.380630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.380641 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.483334 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.483385 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.483401 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.483420 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.483432 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.586436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.586477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.586486 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.586502 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.586511 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.688132 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.688165 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.688173 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.688189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.688198 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.791180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.791221 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.791233 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.791544 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.791566 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.893819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.894254 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.894263 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.894278 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.894286 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.997478 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.997539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.997550 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.997570 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:35 crc kubenswrapper[4690]: I1203 13:10:35.997583 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:35Z","lastTransitionTime":"2025-12-03T13:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.099511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.099559 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.099570 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.099585 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.099596 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.202747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.203199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.203331 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.203459 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.203584 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.263720 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.263940 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:40.263906889 +0000 UTC m=+146.244827332 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.264226 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.264307 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.264384 4690 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.264463 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:11:40.264448022 +0000 UTC m=+146.245368515 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.264610 4690 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.264726 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:11:40.264711499 +0000 UTC m=+146.245631932 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.306521 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.306556 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.306568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.306581 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.306590 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.314257 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.314257 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.314394 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.314455 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.365152 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.365204 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.365356 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.365393 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.365408 4690 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.365457 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:11:40.365441968 +0000 UTC m=+146.346362401 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.365359 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.365499 4690 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.365510 4690 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:10:36 crc kubenswrapper[4690]: E1203 13:10:36.365553 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:11:40.365540481 +0000 UTC m=+146.346460914 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.408967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.409065 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.409086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.409105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.409115 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.511909 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.511947 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.511960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.511977 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.511987 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.614661 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.614703 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.614715 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.614732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.614744 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.717470 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.717499 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.717507 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.717520 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.717530 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.819767 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.819925 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.819942 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.819960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.819973 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.921918 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.921956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.921967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.921980 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:36 crc kubenswrapper[4690]: I1203 13:10:36.921989 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:36Z","lastTransitionTime":"2025-12-03T13:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.024964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.025007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.025016 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.025033 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.025042 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.128045 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.128107 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.128117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.128133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.128142 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.230959 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.230994 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.231003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.231017 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.231026 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.314059 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.314143 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:37 crc kubenswrapper[4690]: E1203 13:10:37.314201 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:37 crc kubenswrapper[4690]: E1203 13:10:37.314267 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.331752 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.331810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.331827 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.331888 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.331901 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: E1203 13:10:37.344204 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.348325 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.348375 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.348387 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.348403 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.348415 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: E1203 13:10:37.360200 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.367172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.367332 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.367356 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.367374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.367384 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: E1203 13:10:37.382144 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.386387 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.386431 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.386440 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.386477 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.386495 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: E1203 13:10:37.400226 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.404712 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.404753 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.404761 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.404776 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.404785 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: E1203 13:10:37.417709 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:37 crc kubenswrapper[4690]: E1203 13:10:37.417917 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.420025 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.420063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.420074 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.420091 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.420103 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.522652 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.522701 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.522712 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.522738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.522749 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.625164 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.625211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.625223 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.625240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.625252 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.728388 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.728438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.728446 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.728463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.728474 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.830362 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.830416 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.830428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.830445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.830457 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.932670 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.932710 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.932724 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.932748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:37 crc kubenswrapper[4690]: I1203 13:10:37.932764 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:37Z","lastTransitionTime":"2025-12-03T13:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.034801 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.034839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.034854 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.034885 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.034899 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.137300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.137327 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.137335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.137350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.137361 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.240331 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.240393 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.240412 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.240432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.240445 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.314470 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.314470 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:38 crc kubenswrapper[4690]: E1203 13:10:38.314615 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:38 crc kubenswrapper[4690]: E1203 13:10:38.314682 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.343473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.343508 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.343519 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.343534 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.343546 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.446857 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.446899 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.446910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.446927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.446939 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.550118 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.550176 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.550190 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.550208 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.550228 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.652585 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.652623 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.652632 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.652647 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.652656 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.755405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.755480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.755495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.755513 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.755535 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.858657 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.858706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.858716 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.858730 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.858740 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.961132 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.961180 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.961191 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.961207 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:38 crc kubenswrapper[4690]: I1203 13:10:38.961221 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:38Z","lastTransitionTime":"2025-12-03T13:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.063700 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.063755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.063773 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.063793 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.063807 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.165821 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.165862 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.165912 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.165930 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.165941 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.269172 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.269225 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.269239 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.269256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.269266 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.313953 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:39 crc kubenswrapper[4690]: E1203 13:10:39.314092 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.313964 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:39 crc kubenswrapper[4690]: E1203 13:10:39.314273 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.372303 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.372357 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.372366 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.372382 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.372402 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.476115 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.476188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.476203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.476222 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.476240 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.578788 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.578840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.578856 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.578908 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.578925 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.682335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.682648 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.682663 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.682680 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.682692 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.784780 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.784819 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.784828 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.784846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.784858 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.887457 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.887497 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.887507 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.887521 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.887530 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.990355 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.990479 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.990493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.990509 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:39 crc kubenswrapper[4690]: I1203 13:10:39.990520 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:39Z","lastTransitionTime":"2025-12-03T13:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.093026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.093071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.093082 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.093100 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.093111 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.195833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.195899 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.195911 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.195930 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.195942 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.299485 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.299527 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.299561 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.299579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.299588 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.314093 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:40 crc kubenswrapper[4690]: E1203 13:10:40.314252 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.314534 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:40 crc kubenswrapper[4690]: E1203 13:10:40.314685 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.315839 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:10:40 crc kubenswrapper[4690]: E1203 13:10:40.316110 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.402041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.402115 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.402129 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.402158 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.402184 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.505842 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.506114 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.506135 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.506162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.506179 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.610395 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.610460 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.610469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.610487 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.610498 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.713652 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.713707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.713717 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.713738 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.713750 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.816381 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.816440 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.816455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.816488 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.816510 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.918821 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.918897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.918910 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.918927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:40 crc kubenswrapper[4690]: I1203 13:10:40.918938 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:40Z","lastTransitionTime":"2025-12-03T13:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.021260 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.021333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.021347 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.021363 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.021375 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.123955 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.123995 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.124007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.124025 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.124039 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.226603 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.226642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.226651 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.226668 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.226698 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.314287 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.314331 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:41 crc kubenswrapper[4690]: E1203 13:10:41.314540 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:41 crc kubenswrapper[4690]: E1203 13:10:41.314670 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.329917 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.329948 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.329957 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.329975 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.329987 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.432103 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.432157 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.432177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.432202 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.432216 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.535630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.535671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.535682 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.535699 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.535711 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.638389 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.638442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.638455 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.638476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.638488 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.740890 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.740933 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.740942 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.740956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.740965 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.843708 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.843790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.843802 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.843822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.843832 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.946091 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.946137 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.946149 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.946169 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:41 crc kubenswrapper[4690]: I1203 13:10:41.946264 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:41Z","lastTransitionTime":"2025-12-03T13:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.048456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.048500 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.048512 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.048527 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.048537 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.151242 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.151279 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.151288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.151315 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.151325 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.257916 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.257968 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.257981 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.258000 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.258015 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.314543 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.314682 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:42 crc kubenswrapper[4690]: E1203 13:10:42.314792 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:42 crc kubenswrapper[4690]: E1203 13:10:42.314906 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.360739 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.361141 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.361216 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.361279 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.361339 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.464518 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.464814 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.465016 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.465130 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.465159 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.567617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.567659 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.567670 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.567685 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.567695 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.670612 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.670664 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.670678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.670694 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.670707 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.773270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.773317 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.773328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.773347 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.773358 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.876034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.876077 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.876085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.876100 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.876109 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.978747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.978787 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.978795 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.978808 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:42 crc kubenswrapper[4690]: I1203 13:10:42.978818 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:42Z","lastTransitionTime":"2025-12-03T13:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.081628 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.081668 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.081679 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.081693 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.081702 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.183995 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.184037 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.184049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.184066 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.184077 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.286940 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.286998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.287015 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.287039 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.287058 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.314521 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.314584 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:43 crc kubenswrapper[4690]: E1203 13:10:43.314683 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:43 crc kubenswrapper[4690]: E1203 13:10:43.314918 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.390111 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.390150 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.390161 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.390177 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.390188 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.493289 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.493334 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.493348 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.493367 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.493380 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.596449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.596491 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.596503 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.596520 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.596533 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.699175 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.699215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.699227 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.699244 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.699255 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.802354 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.802398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.802418 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.802439 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.802453 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.904197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.904275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.904288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.904307 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:43 crc kubenswrapper[4690]: I1203 13:10:43.904320 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:43Z","lastTransitionTime":"2025-12-03T13:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.006435 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.006489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.006501 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.006518 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.006527 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.109328 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.109380 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.109397 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.109420 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.109439 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.212859 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.212941 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.212956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.212979 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.212997 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.314476 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.314567 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:44 crc kubenswrapper[4690]: E1203 13:10:44.314712 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:44 crc kubenswrapper[4690]: E1203 13:10:44.314954 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.318370 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.318427 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.318444 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.318470 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.318487 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.331485 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.342708 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.355205 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.367329 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.380258 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.428755 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afdda9eb-0b52-4293-a70f-7ee5a507df8b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9a2357cbe74fa94cec46032822a70d9124c382e568d434ec7dfc64167aa8aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.430057 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.430079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.430088 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.430105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.430115 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.452857 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.467427 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.478485 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.493117 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.510493 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.525429 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.531821 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.531892 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.531902 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.531923 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.531933 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.536361 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.556472 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.570973 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.582797 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.593776 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.611588 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:28Z\\\",\\\"message\\\":\\\":10:27.998491 6669 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:10:27.996511 6669 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:10:27.998477 6669 services_controller.go:453] Built service openshift-network-console/networking-console-plugin template LB for network=default: []services.LB{}\\\\nF1203 13:10:27.998530 6669 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start nod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:10:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.621103 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.633606 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.633630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.633639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.633652 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.633661 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.736965 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.737007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.737018 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.737034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.737045 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.839079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.839714 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.839966 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.840070 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.840163 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.942792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.943150 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.943278 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.943365 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:44 crc kubenswrapper[4690]: I1203 13:10:44.943429 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:44Z","lastTransitionTime":"2025-12-03T13:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.045725 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.046100 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.046211 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.046309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.046392 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.149014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.149087 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.149098 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.149114 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.149124 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.251692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.251755 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.251770 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.251789 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.251802 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.313996 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.314062 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:45 crc kubenswrapper[4690]: E1203 13:10:45.314392 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:45 crc kubenswrapper[4690]: E1203 13:10:45.314553 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.354220 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.354554 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.354648 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.354748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.354849 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.457790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.458101 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.458189 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.458289 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.458387 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.560527 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.560562 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.560573 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.560590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.560603 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.665756 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.665821 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.665846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.665944 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.665972 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.768976 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.769006 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.769017 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.769034 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.769046 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.872574 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.872634 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.872655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.872683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.872704 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.974963 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.975012 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.975029 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.975047 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:45 crc kubenswrapper[4690]: I1203 13:10:45.975067 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:45Z","lastTransitionTime":"2025-12-03T13:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.078424 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.078807 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.078929 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.079027 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.079151 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.180773 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.180801 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.180810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.180823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.180831 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.283539 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.283579 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.283590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.283608 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.283620 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.314359 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.314383 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:46 crc kubenswrapper[4690]: E1203 13:10:46.314532 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:46 crc kubenswrapper[4690]: E1203 13:10:46.314640 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.386059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.386096 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.386105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.386119 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.386129 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.488621 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.488696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.488710 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.488726 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.488736 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.591427 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.591484 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.591500 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.591522 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.591545 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.694255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.694323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.694345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.694365 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.694378 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.797363 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.797419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.797438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.797462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.797479 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.900517 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.900567 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.900577 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.900593 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:46 crc kubenswrapper[4690]: I1203 13:10:46.900601 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:46Z","lastTransitionTime":"2025-12-03T13:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.003408 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.003449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.003476 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.003495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.003506 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.106954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.107401 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.107632 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.107843 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.108048 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.210928 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.210978 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.210992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.211012 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.211024 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.313222 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.313310 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.313329 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.313361 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.313381 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.313785 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.313796 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:47 crc kubenswrapper[4690]: E1203 13:10:47.314555 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:47 crc kubenswrapper[4690]: E1203 13:10:47.314606 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.416215 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.416242 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.416250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.416265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.416326 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.518884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.518947 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.518960 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.518979 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.518990 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.620954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.621000 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.621010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.621028 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.621037 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.723781 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.723909 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.723926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.723951 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.723965 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.742437 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.742493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.742506 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.742526 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.742541 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: E1203 13:10:47.756286 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.760418 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.760445 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.760452 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.760465 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.760499 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: E1203 13:10:47.773006 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.815952 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.816009 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.816022 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.816042 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.816055 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: E1203 13:10:47.830662 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.839449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.839501 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.839510 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.839523 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.839534 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: E1203 13:10:47.852608 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.856850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.856919 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.856933 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.856954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.856969 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: E1203 13:10:47.869367 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:47Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:47 crc kubenswrapper[4690]: E1203 13:10:47.869562 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.871369 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.871405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.871414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.871434 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.871444 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.974442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.974471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.974480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.974493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:47 crc kubenswrapper[4690]: I1203 13:10:47.974502 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:47Z","lastTransitionTime":"2025-12-03T13:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.077814 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.078133 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.078162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.078192 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.078215 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.181213 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.181288 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.181306 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.181323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.181334 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.283490 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.283550 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.283563 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.283582 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.283627 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.314284 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.314291 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:48 crc kubenswrapper[4690]: E1203 13:10:48.314443 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:48 crc kubenswrapper[4690]: E1203 13:10:48.314579 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.386046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.386083 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.386094 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.386110 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.386121 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.488531 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.488584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.488594 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.488612 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.488622 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.590707 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.590741 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.590749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.590763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.590772 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.693574 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.693627 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.693639 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.693656 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.693668 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.798195 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.798300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.798831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.798946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.799268 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.901980 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.902020 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.902028 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.902045 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:48 crc kubenswrapper[4690]: I1203 13:10:48.902054 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:48Z","lastTransitionTime":"2025-12-03T13:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.004401 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.004438 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.004447 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.004462 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.004472 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.107834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.107934 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.107954 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.107984 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.108008 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.210617 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.210667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.210678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.210696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.210710 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.313574 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.313667 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.313692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.313689 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.313729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.313754 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: E1203 13:10:49.313844 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.313932 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:49 crc kubenswrapper[4690]: E1203 13:10:49.314027 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.416589 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.416634 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.416646 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.416663 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.416676 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.519277 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.519326 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.519340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.519359 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.519376 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.622006 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.622050 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.622059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.622075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.622086 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.724642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.724697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.724713 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.724732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.724746 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.827740 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.827857 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.827891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.827916 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.827933 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.930676 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.930723 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.930740 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.930762 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:49 crc kubenswrapper[4690]: I1203 13:10:49.930778 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:49Z","lastTransitionTime":"2025-12-03T13:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.033051 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.033094 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.033105 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.033124 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.033137 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.135466 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.135522 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.135531 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.135548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.135562 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.237678 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.237719 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.237732 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.237747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.237756 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.314017 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.314092 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:50 crc kubenswrapper[4690]: E1203 13:10:50.314202 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:50 crc kubenswrapper[4690]: E1203 13:10:50.314277 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.340806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.340851 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.340860 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.340897 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.340909 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.443907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.443979 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.443997 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.444023 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.444050 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.546908 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.546958 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.546970 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.546988 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.547001 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.649280 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.649318 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.649327 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.649340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.649349 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.751796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.751824 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.751833 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.751846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.751856 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.854914 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.854966 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.854977 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.854993 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.855003 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.957206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.957235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.957242 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.957255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:50 crc kubenswrapper[4690]: I1203 13:10:50.957263 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:50Z","lastTransitionTime":"2025-12-03T13:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.059839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.059972 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.059998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.060029 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.060051 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.163433 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.163525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.163548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.163584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.163607 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.273428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.273470 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.273479 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.273497 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.273506 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.314566 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.314680 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:51 crc kubenswrapper[4690]: E1203 13:10:51.314774 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:51 crc kubenswrapper[4690]: E1203 13:10:51.314879 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.376794 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.376837 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.376846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.376861 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.376888 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.480729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.480771 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.480781 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.480796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.480808 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.583792 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.583844 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.583855 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.583903 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.583915 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.687685 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.688020 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.688039 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.688066 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.688164 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.791341 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.791377 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.791385 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.791398 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.791408 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.894366 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.894449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.894485 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.894515 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.894538 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.997840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.997929 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.997950 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.997974 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:51 crc kubenswrapper[4690]: I1203 13:10:51.997995 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:51Z","lastTransitionTime":"2025-12-03T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.030851 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:52 crc kubenswrapper[4690]: E1203 13:10:52.031188 4690 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:10:52 crc kubenswrapper[4690]: E1203 13:10:52.031314 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs podName:abab6c63-91d1-4f36-9be9-ae3b1a43742e nodeName:}" failed. No retries permitted until 2025-12-03 13:11:56.031281892 +0000 UTC m=+162.012202365 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs") pod "network-metrics-daemon-mbbpr" (UID: "abab6c63-91d1-4f36-9be9-ae3b1a43742e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.101157 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.101199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.101212 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.101230 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.101242 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.204764 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.205192 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.205283 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.205389 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.205449 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.307990 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.308117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.308148 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.308182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.308206 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.315536 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:52 crc kubenswrapper[4690]: E1203 13:10:52.315685 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.316183 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:52 crc kubenswrapper[4690]: E1203 13:10:52.316350 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.411341 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.411403 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.411413 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.411431 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.411441 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.514290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.514339 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.514347 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.514362 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.514372 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.617257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.617324 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.617342 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.617374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.617396 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.720696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.720754 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.720763 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.720782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.720792 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.824335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.824407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.824431 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.824469 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.824496 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.927673 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.927759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.927801 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.927840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:52 crc kubenswrapper[4690]: I1203 13:10:52.927908 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:52Z","lastTransitionTime":"2025-12-03T13:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.031602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.031683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.031711 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.031747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.031774 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.133993 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.134060 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.134074 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.134090 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.134103 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.237766 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.237823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.237835 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.237852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.237864 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.313902 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.313996 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:53 crc kubenswrapper[4690]: E1203 13:10:53.314118 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:53 crc kubenswrapper[4690]: E1203 13:10:53.314256 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.315075 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:10:53 crc kubenswrapper[4690]: E1203 13:10:53.315276 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.341085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.341142 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.341155 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.341174 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.341186 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.444217 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.444270 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.444281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.444300 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.444311 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.547681 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.547739 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.547749 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.547769 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.547781 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.650371 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.650425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.650439 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.650458 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.650471 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.753706 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.753779 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.753793 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.753818 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.753833 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.857030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.857095 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.857104 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.857125 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.857138 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.960436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.960508 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.960523 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.960542 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:53 crc kubenswrapper[4690]: I1203 13:10:53.960554 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:53Z","lastTransitionTime":"2025-12-03T13:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.063117 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.063170 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.063179 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.063195 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.063208 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.165564 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.165616 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.165633 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.165655 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.165670 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.268425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.268458 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.268472 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.268487 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.268499 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.314580 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:54 crc kubenswrapper[4690]: E1203 13:10:54.314769 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.315078 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:54 crc kubenswrapper[4690]: E1203 13:10:54.315161 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.328052 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tf2cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf580211-e03a-444e-98b8-c12f99091b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:20Z\\\",\\\"message\\\":\\\"2025-12-03T13:09:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079\\\\n2025-12-03T13:09:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61660d93-f48d-4882-8cd3-84581b8d1079 to /host/opt/cni/bin/\\\\n2025-12-03T13:09:35Z [verbose] multus-daemon started\\\\n2025-12-03T13:09:35Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:10:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwc65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tf2cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.339327 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f1e382c-408f-4b5c-a51f-2f90770af7fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66e039d5a2a5a5fe27d28b461b2be8e81a4930f97db9de3a71151956d1e67fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://affb549801c485ae7da2a7cde0a3a2bbd5e024bdd7db110de213dfbcde713dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5h9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2hc7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.350563 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.369132 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dthlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545e942f-5352-4a89-8e75-17762d512364\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://398fb79803afb73211795c434968f81f49d0c4672ba336256fd08d4f0d988312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8a3fad2719c921bf71269e1ca228234de444d2fcb957ecd71935e0c1df3c7ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a893c5e11ac6c0211788e4c6c4736f006cbe84f62cfc57803e194068431310c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d91a872282929078a3a8d66b13e2e5ad0edced7ad486523176f983efeadb0cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e61aac0f8d4367524292153236c88df14f29bba1e4a9405026e594442636f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1507d2599b2f6452b03c278278c90143d50b7445b80e3a7ae6884bca2bb3e0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://438bace87e4a9225069bd6763302a714632c08ea3fad8677c42bdf219507c928\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxp5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dthlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.370917 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.370945 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.370955 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.370971 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.370982 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.382021 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.395717 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e23e6be9-62d7-4b0a-a073-52e9fc3a8723\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c0137296b91fc78f7eba3dfd02d5e4c787bdfc8212153f8c4a7048831f40132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5bd7d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5fc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.416159 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8490fb8-480e-4d81-82dd-4411b1f29278\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:10:28Z\\\",\\\"message\\\":\\\":10:27.998491 6669 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:10:27.996511 6669 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:10:27.998477 6669 services_controller.go:453] Built service openshift-network-console/networking-console-plugin template LB for network=default: []services.LB{}\\\\nF1203 13:10:27.998530 6669 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start nod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:10:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7hgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vmddn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.425472 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gkl8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f190c94-af05-4d67-8459-c02f4ecc81b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d66015aed66cf797893d83aba9e007ac4279a3b6527f3d4ac1f6cb620070d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hpd2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gkl8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.443588 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4764c033-44c3-464a-a4f8-356a22400d3c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4068c039c995a749fa8020a48ae09322d69f3749f2e2315665093aa3ae1778a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67f4fdc5ea810e9ab4fe15c7df5066fff9354be1336f261bb6bc858a52d7b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b294be17a5ef3cc25cb4e9e7a603b323598a0ea4cb18e764dd5722f7f182e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3fd75f27ad6a0738b8cfc0a77f4159a89fcaf16ba5479a5a09f6addf66e1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18d85a46c17c039adceb5b65f4137b95ad3a4790934db3d280346ff702a9c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89f714168e7d18cadaf54b4a62462294828967fd4014a122a17896a27d37850\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96c8f13e08fcfd98f400f2a84afb5f5f3ac99a1c9777c09021fefe5d7d822ada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://247dff14c131cac215cfc1ed8c28f31c3264736a921addf7df751fdf5d3ee587\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.454896 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf4934bfc3d49358e91e9d99b826045f949b16d24d2b109036801be9f6595d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.466379 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qh4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcf22e98-c7f6-4eb3-9c13-60f1953a9271\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9129473be08bca04abd3ee2c97cb509a1903e98659e8da2a06736cf6b0802b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tpngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qh4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.473611 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.473653 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.473669 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.473685 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.473697 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.477902 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf72335-5cc2-4b71-ba45-5646e0e44147\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2042389a89dcde3a710fea6b8297876e69712e4fb7fc4344b15c5bee9423352e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af082211b200b04416f63bf381e56cb850cd3ba36ca8d2483756c27b50609eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f6f36d150886633848803189685dd8084889d5ce9fbb0779ba36f0afa43240\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://884a6145f85fa670b64e94fcc9c9af80e56d1c6e569a9d8be3309fd7c107214f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.488942 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1593a0ee1bee46ffb72619de3adb64986fd1e2e52c574ad201a9a52f77d84b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.498450 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afdda9eb-0b52-4293-a70f-7ee5a507df8b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9a2357cbe74fa94cec46032822a70d9124c382e568d434ec7dfc64167aa8aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://626838dfa99af9ad8da6a9f4f92eb9b8fc0f281eab80bed500f2738cf76c70d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.510344 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.523415 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706472baa0445759d66ecc90c972da51fafdbad315c21d0792db2c96a757848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4884b398cf14c5623e0e5e46fe7e0df41423e75b148f216e402db7b1173a3ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.533212 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abab6c63-91d1-4f36-9be9-ae3b1a43742e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p92gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mbbpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.549795 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e2e662-ec6e-499e-a68f-b57d17aa218e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:09:32Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:09:26.887044 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:09:26.894583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-923344266/tls.crt::/tmp/serving-cert-923344266/tls.key\\\\\\\"\\\\nI1203 13:09:32.239060 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:09:32.242197 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:09:32.242225 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:09:32.242249 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:09:32.242255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:09:32.247961 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 13:09:32.247982 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:09:32.247990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:09:32.247994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:09:32.247996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:09:32.247999 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 13:09:32.247997 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 13:09:32.252420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.562782 4690 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e799f5-8596-4229-be2a-cbb5ac744b24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://477caf47b4b4f9cb45b09be15dbbc6d9ddff056f1087d2c2dc8278dc8b84681d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f288f4ed094058ade247b98e3c97c79e81bd3d809bf1e47d2fc61bdc9ebefa90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc536e6240a9906e7d8b283000a80fed723e62efe6e8379ba0a155767128bb52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:09:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:54Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.576989 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.577036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.577046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.577061 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.577071 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.679264 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.679345 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.679358 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.679378 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.679395 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.782634 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.782674 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.782683 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.782697 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.782709 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.885203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.885255 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.885265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.885282 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.885296 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.989101 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.989162 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.989182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.989209 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:54 crc kubenswrapper[4690]: I1203 13:10:54.989232 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:54Z","lastTransitionTime":"2025-12-03T13:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.092374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.092428 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.092437 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.092454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.092466 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.195532 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.195564 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.195572 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.195587 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.195596 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.298357 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.298411 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.298425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.298447 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.298459 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.313979 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.314024 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:55 crc kubenswrapper[4690]: E1203 13:10:55.314459 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:55 crc kubenswrapper[4690]: E1203 13:10:55.314530 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.401031 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.401074 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.401085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.401104 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.401116 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.503414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.503449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.503456 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.503470 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.503478 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.606127 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.606164 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.606171 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.606187 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.606198 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.707980 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.708021 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.708032 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.708046 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.708058 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.810761 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.810805 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.810817 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.810834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.810847 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.913518 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.913557 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.913569 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.913584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:55 crc kubenswrapper[4690]: I1203 13:10:55.913597 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:55Z","lastTransitionTime":"2025-12-03T13:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.016350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.016396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.016407 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.016422 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.016432 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.119333 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.119449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.119468 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.119490 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.119507 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.222151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.222192 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.222207 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.222224 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.222237 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.314780 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.314927 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:56 crc kubenswrapper[4690]: E1203 13:10:56.315061 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:56 crc kubenswrapper[4690]: E1203 13:10:56.315192 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.324240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.324278 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.324296 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.324313 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.324325 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.426200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.426241 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.426249 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.426263 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.426274 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.529545 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.529598 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.529611 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.529629 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.529643 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.632785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.632843 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.632852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.632891 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.632905 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.736778 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.736822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.736830 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.736845 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.736860 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.839444 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.839772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.839798 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.839823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.839845 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.942353 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.942396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.942408 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.942423 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:56 crc kubenswrapper[4690]: I1203 13:10:56.942435 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:56Z","lastTransitionTime":"2025-12-03T13:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.045570 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.045633 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.045658 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.045690 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.045713 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.148443 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.148506 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.148523 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.148550 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.148570 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.251730 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.251790 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.251844 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.251860 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.251953 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.314172 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.314212 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:57 crc kubenswrapper[4690]: E1203 13:10:57.314373 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:57 crc kubenswrapper[4690]: E1203 13:10:57.314516 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.355320 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.355555 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.355568 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.355585 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.355600 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.458236 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.458281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.458293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.458308 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.458319 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.561693 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.561810 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.561826 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.561850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.561910 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.665030 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.665092 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.665108 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.665134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.665152 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.768374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.768434 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.768449 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.768471 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.768491 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.871520 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.872038 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.872237 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.872425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.872624 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.976035 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.976092 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.976104 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.976125 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:57 crc kubenswrapper[4690]: I1203 13:10:57.976140 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:57Z","lastTransitionTime":"2025-12-03T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.025148 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.025239 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.025251 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.025273 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.025292 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: E1203 13:10:58.040794 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.045003 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.045048 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.045063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.045086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.045103 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: E1203 13:10:58.058264 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.062226 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.062374 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.062473 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.062551 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.062632 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: E1203 13:10:58.076267 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.080494 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.080527 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.080538 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.080553 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.080563 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: E1203 13:10:58.094712 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.098834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.098922 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.098938 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.098963 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.098979 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: E1203 13:10:58.113072 4690 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:10:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95a051aa-8ddb-4533-bbde-548417042fa3\\\",\\\"systemUUID\\\":\\\"c6db3655-bd43-479c-8c13-6f91f454fdd0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:10:58Z is after 2025-08-24T17:21:41Z" Dec 03 13:10:58 crc kubenswrapper[4690]: E1203 13:10:58.113190 4690 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.114943 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.114991 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.115006 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.115026 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.115039 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.217774 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.217825 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.217837 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.217858 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.217896 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.314402 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:10:58 crc kubenswrapper[4690]: E1203 13:10:58.314778 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.314413 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:10:58 crc kubenswrapper[4690]: E1203 13:10:58.315086 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.320001 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.320054 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.320065 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.320079 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.320089 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.422956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.423049 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.423063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.423085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.423099 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.525552 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.525597 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.525609 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.525630 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.525643 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.628535 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.628967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.628986 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.629009 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.629029 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.732520 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.732561 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.732571 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.732586 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.732597 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.836062 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.836115 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.836128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.836146 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.836169 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.938996 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.939051 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.939065 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.939085 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:58 crc kubenswrapper[4690]: I1203 13:10:58.939098 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:58Z","lastTransitionTime":"2025-12-03T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.041691 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.041735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.041748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.041764 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.041775 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.144939 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.145004 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.145027 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.145070 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.145091 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.247769 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.247813 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.247824 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.247841 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.247854 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.313994 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.314067 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:10:59 crc kubenswrapper[4690]: E1203 13:10:59.314227 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:10:59 crc kubenswrapper[4690]: E1203 13:10:59.314369 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.350014 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.350041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.350065 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.350077 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.350088 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.453196 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.453250 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.453267 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.453290 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.453307 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.556279 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.556346 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.556362 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.556386 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.556402 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.658551 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.658595 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.658608 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.658625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.658635 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.760526 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.760561 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.760573 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.760589 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.760606 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.862956 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.863315 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.863409 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.863493 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.863581 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.966806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.967291 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.967516 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.967728 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:10:59 crc kubenswrapper[4690]: I1203 13:10:59.967972 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:10:59Z","lastTransitionTime":"2025-12-03T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.069884 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.069917 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.069926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.069941 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.069951 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.173287 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.173323 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.173335 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.173350 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.173359 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.276063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.276396 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.276511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.276615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.276711 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.314446 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.314526 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:00 crc kubenswrapper[4690]: E1203 13:11:00.315058 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:00 crc kubenswrapper[4690]: E1203 13:11:00.315248 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.378850 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.378929 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.378942 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.378964 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.378981 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.481392 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.481457 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.481480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.481509 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.481531 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.585517 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.585597 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.585615 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.585642 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.585666 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.687918 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.687990 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.688010 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.688036 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.688053 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.791006 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.791075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.791092 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.791121 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.791140 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.893343 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.893382 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.893391 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.893405 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.893415 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.996489 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.996531 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.996541 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.996561 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:00 crc kubenswrapper[4690]: I1203 13:11:00.996580 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:00Z","lastTransitionTime":"2025-12-03T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.100797 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.100907 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.100927 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.100953 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.100970 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.204672 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.204729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.204747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.204772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.204791 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.308041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.308086 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.308199 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.308219 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.308231 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.313963 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.313999 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:01 crc kubenswrapper[4690]: E1203 13:11:01.314132 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:01 crc kubenswrapper[4690]: E1203 13:11:01.314390 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.411459 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.411540 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.411560 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.411584 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.411603 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.515206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.515256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.515273 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.515298 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.515315 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.617443 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.617495 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.617511 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.617531 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.617545 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.719602 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.719646 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.719657 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.719675 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.719687 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.822750 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.822806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.822854 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.822900 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.822942 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.926123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.926170 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.926182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.926200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:01 crc kubenswrapper[4690]: I1203 13:11:01.926211 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:01Z","lastTransitionTime":"2025-12-03T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.028446 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.028480 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.028490 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.028509 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.028522 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.131126 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.131176 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.131188 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.131207 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.131219 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.233782 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.233820 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.233831 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.233846 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.233905 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.313770 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.313903 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:02 crc kubenswrapper[4690]: E1203 13:11:02.313940 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:02 crc kubenswrapper[4690]: E1203 13:11:02.314065 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.336696 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.336735 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.336745 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.336759 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.336768 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.438851 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.438923 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.438946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.438967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.438977 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.541157 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.541196 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.541209 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.541226 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.541240 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.644051 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.644100 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.644111 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.644129 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.644141 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.746796 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.746905 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.746929 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.746952 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.746966 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.849414 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.849454 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.849467 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.849483 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.849494 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.951465 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.951525 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.951546 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.951564 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:02 crc kubenswrapper[4690]: I1203 13:11:02.951577 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:02Z","lastTransitionTime":"2025-12-03T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.053298 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.053347 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.053359 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.053378 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.053392 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.155340 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.155389 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.155404 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.155426 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.155439 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.257784 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.257823 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.257834 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.257849 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.257859 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.314367 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.314477 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:03 crc kubenswrapper[4690]: E1203 13:11:03.314519 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:03 crc kubenswrapper[4690]: E1203 13:11:03.314650 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.360262 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.360311 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.360322 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.360338 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.360350 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.462625 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.462666 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.462677 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.462692 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.462703 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.565309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.565354 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.565373 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.565391 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.565400 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.667383 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.667425 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.667436 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.667452 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.667462 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.770443 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.770499 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.770510 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.770534 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.770543 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.873134 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.873170 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.873182 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.873197 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.873209 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.975710 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.975747 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.975757 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.975772 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:03 crc kubenswrapper[4690]: I1203 13:11:03.975783 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:03Z","lastTransitionTime":"2025-12-03T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.077748 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.077788 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.077801 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.077817 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.077828 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.180256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.180341 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.180354 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.180377 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.180385 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.282774 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.282817 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.282826 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.282839 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.282847 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.314326 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.314324 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:04 crc kubenswrapper[4690]: E1203 13:11:04.314471 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:04 crc kubenswrapper[4690]: E1203 13:11:04.314612 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.343758 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.343734925 podStartE2EDuration="57.343734925s" podCreationTimestamp="2025-12-03 13:10:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.332302971 +0000 UTC m=+110.313223434" watchObservedRunningTime="2025-12-03 13:11:04.343734925 +0000 UTC m=+110.324655358" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.352253 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-qh4c5" podStartSLOduration=91.352234786 podStartE2EDuration="1m31.352234786s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.351766564 +0000 UTC m=+110.332686997" watchObservedRunningTime="2025-12-03 13:11:04.352234786 +0000 UTC m=+110.333155219" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.379525 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.379506532 podStartE2EDuration="1m32.379506532s" podCreationTimestamp="2025-12-03 13:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.37940387 +0000 UTC m=+110.360324313" watchObservedRunningTime="2025-12-03 13:11:04.379506532 +0000 UTC m=+110.360426965" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.388522 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.388806 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.388894 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.388967 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.389028 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.395690 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=90.395674663 podStartE2EDuration="1m30.395674663s" podCreationTimestamp="2025-12-03 13:09:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.395070488 +0000 UTC m=+110.375990921" watchObservedRunningTime="2025-12-03 13:11:04.395674663 +0000 UTC m=+110.376595096" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.406011 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=38.40599352 podStartE2EDuration="38.40599352s" podCreationTimestamp="2025-12-03 13:10:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.405793045 +0000 UTC m=+110.386713488" watchObservedRunningTime="2025-12-03 13:11:04.40599352 +0000 UTC m=+110.386913953" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.457399 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-dthlv" podStartSLOduration=91.457377914 podStartE2EDuration="1m31.457377914s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.457147809 +0000 UTC m=+110.438068262" watchObservedRunningTime="2025-12-03 13:11:04.457377914 +0000 UTC m=+110.438298357" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.470591 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-tf2cb" podStartSLOduration=91.470573482 podStartE2EDuration="1m31.470573482s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.470017028 +0000 UTC m=+110.450937481" watchObservedRunningTime="2025-12-03 13:11:04.470573482 +0000 UTC m=+110.451493915" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.481945 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2hc7d" podStartSLOduration=91.481924444 podStartE2EDuration="1m31.481924444s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.481618056 +0000 UTC m=+110.462538489" watchObservedRunningTime="2025-12-03 13:11:04.481924444 +0000 UTC m=+110.462844877" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.491063 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.491371 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.491443 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.491509 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.491568 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.519161 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=92.519140517 podStartE2EDuration="1m32.519140517s" podCreationTimestamp="2025-12-03 13:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.518660505 +0000 UTC m=+110.499580948" watchObservedRunningTime="2025-12-03 13:11:04.519140517 +0000 UTC m=+110.500060950" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.519533 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-gkl8h" podStartSLOduration=91.519527977 podStartE2EDuration="1m31.519527977s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.492173088 +0000 UTC m=+110.473093541" watchObservedRunningTime="2025-12-03 13:11:04.519527977 +0000 UTC m=+110.500448410" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.561419 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podStartSLOduration=91.561399136 podStartE2EDuration="1m31.561399136s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:04.560807631 +0000 UTC m=+110.541728054" watchObservedRunningTime="2025-12-03 13:11:04.561399136 +0000 UTC m=+110.542319559" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.594545 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.594580 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.594589 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.594603 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.594614 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.697992 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.698033 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.698044 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.698059 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.698071 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.800115 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.800153 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.800163 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.800179 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.800190 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.902110 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.902154 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.902163 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.902214 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:04 crc kubenswrapper[4690]: I1203 13:11:04.902227 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:04Z","lastTransitionTime":"2025-12-03T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.005200 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.005235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.005243 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.005256 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.005265 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.107804 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.107840 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.107857 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.107893 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.107903 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.210432 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.210463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.210472 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.210486 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.210496 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.313219 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.313265 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.313277 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.313293 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.313303 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.313834 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:05 crc kubenswrapper[4690]: E1203 13:11:05.313944 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.313846 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:05 crc kubenswrapper[4690]: E1203 13:11:05.314011 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.415882 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.415926 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.415934 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.415949 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.415958 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.518583 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.518628 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.518636 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.518649 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.518659 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.621071 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.621107 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.621116 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.621130 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.621139 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.723573 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.723604 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.723612 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.723626 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.723635 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.825417 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.825463 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.825472 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.825487 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.825498 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.928206 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.928257 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.928275 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.928292 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:05 crc kubenswrapper[4690]: I1203 13:11:05.928305 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:05Z","lastTransitionTime":"2025-12-03T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.031285 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.031319 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.031326 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.031339 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.031348 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.134354 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.134401 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.134419 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.134442 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.134459 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.237021 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.237053 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.237062 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.237075 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.237086 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.314332 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.314400 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:06 crc kubenswrapper[4690]: E1203 13:11:06.314561 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:06 crc kubenswrapper[4690]: E1203 13:11:06.314698 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.338921 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.338998 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.339021 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.339050 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.339074 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.441822 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.441883 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.441893 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.441909 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.441922 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.544999 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.545076 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.545099 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.545128 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.545150 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.647712 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.647756 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.647766 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.647785 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.647799 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.749830 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.749894 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.749912 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.749928 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.749938 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.852515 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.852611 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.852641 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.852671 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.852694 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.911025 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/1.log" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.911525 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/0.log" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.911565 4690 generic.go:334] "Generic (PLEG): container finished" podID="bf580211-e03a-444e-98b8-c12f99091b80" containerID="f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6" exitCode=1 Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.911595 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tf2cb" event={"ID":"bf580211-e03a-444e-98b8-c12f99091b80","Type":"ContainerDied","Data":"f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6"} Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.911630 4690 scope.go:117] "RemoveContainer" containerID="7a79c6e00560f0d88d569d9881047bf63a554ac18c22b7ad014b01ba8755621a" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.912241 4690 scope.go:117] "RemoveContainer" containerID="f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6" Dec 03 13:11:06 crc kubenswrapper[4690]: E1203 13:11:06.912536 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-tf2cb_openshift-multus(bf580211-e03a-444e-98b8-c12f99091b80)\"" pod="openshift-multus/multus-tf2cb" podUID="bf580211-e03a-444e-98b8-c12f99091b80" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.955109 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.955159 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.955174 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.955195 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:06 crc kubenswrapper[4690]: I1203 13:11:06.955215 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:06Z","lastTransitionTime":"2025-12-03T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.058151 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.058203 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.058218 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.058240 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.058255 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.161805 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.161898 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.161922 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.161949 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.161962 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.264235 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.264281 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.264292 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.264309 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.264320 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.314208 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.314247 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:07 crc kubenswrapper[4690]: E1203 13:11:07.314373 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:07 crc kubenswrapper[4690]: E1203 13:11:07.314759 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.315075 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:11:07 crc kubenswrapper[4690]: E1203 13:11:07.315236 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vmddn_openshift-ovn-kubernetes(c8490fb8-480e-4d81-82dd-4411b1f29278)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.366000 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.366041 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.366051 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.366067 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.366077 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.467940 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.467980 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.467991 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.468007 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.468017 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.570852 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.570931 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.570946 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.570969 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.570984 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.673483 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.673855 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.674025 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.674123 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.674186 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.776560 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.777150 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.777229 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.777299 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.777355 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.879536 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.879576 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.879588 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.879608 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.879621 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.915897 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/1.log" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.982505 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.982548 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.982561 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.982576 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:07 crc kubenswrapper[4690]: I1203 13:11:07.982587 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:07Z","lastTransitionTime":"2025-12-03T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.084881 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.084936 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.084947 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.084961 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.084971 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:08Z","lastTransitionTime":"2025-12-03T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.186531 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.186564 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.186574 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.186590 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.186602 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:08Z","lastTransitionTime":"2025-12-03T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.255324 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.255361 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.255368 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.255383 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.255393 4690 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:11:08Z","lastTransitionTime":"2025-12-03T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.298622 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q"] Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.299506 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.301257 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.302128 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.302367 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.304431 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.316372 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:08 crc kubenswrapper[4690]: E1203 13:11:08.316536 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.316936 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:08 crc kubenswrapper[4690]: E1203 13:11:08.317052 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.402260 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66ba3c0d-9612-4a69-b300-fd60bc949dbe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.402359 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66ba3c0d-9612-4a69-b300-fd60bc949dbe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.402412 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66ba3c0d-9612-4a69-b300-fd60bc949dbe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.402478 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66ba3c0d-9612-4a69-b300-fd60bc949dbe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.402610 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66ba3c0d-9612-4a69-b300-fd60bc949dbe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.503377 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66ba3c0d-9612-4a69-b300-fd60bc949dbe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.503536 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66ba3c0d-9612-4a69-b300-fd60bc949dbe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.503630 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66ba3c0d-9612-4a69-b300-fd60bc949dbe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.503688 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66ba3c0d-9612-4a69-b300-fd60bc949dbe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.503702 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66ba3c0d-9612-4a69-b300-fd60bc949dbe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.503720 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66ba3c0d-9612-4a69-b300-fd60bc949dbe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.503822 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66ba3c0d-9612-4a69-b300-fd60bc949dbe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.504909 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66ba3c0d-9612-4a69-b300-fd60bc949dbe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.512941 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66ba3c0d-9612-4a69-b300-fd60bc949dbe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.524184 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66ba3c0d-9612-4a69-b300-fd60bc949dbe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-flm4q\" (UID: \"66ba3c0d-9612-4a69-b300-fd60bc949dbe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.614349 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.920398 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" event={"ID":"66ba3c0d-9612-4a69-b300-fd60bc949dbe","Type":"ContainerStarted","Data":"a92a62f451b2bcfd9c472668fc505e20529f8828dc42703ac7f72431c8cd7e15"} Dec 03 13:11:08 crc kubenswrapper[4690]: I1203 13:11:08.920930 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" event={"ID":"66ba3c0d-9612-4a69-b300-fd60bc949dbe","Type":"ContainerStarted","Data":"819adf81c9fc625905e06d3ca83320becce3e0ac6b598a37661a8707e81a54e5"} Dec 03 13:11:09 crc kubenswrapper[4690]: I1203 13:11:09.313838 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:09 crc kubenswrapper[4690]: I1203 13:11:09.313944 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:09 crc kubenswrapper[4690]: E1203 13:11:09.314043 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:09 crc kubenswrapper[4690]: E1203 13:11:09.314199 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:10 crc kubenswrapper[4690]: I1203 13:11:10.314329 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:10 crc kubenswrapper[4690]: E1203 13:11:10.314516 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:10 crc kubenswrapper[4690]: I1203 13:11:10.314329 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:10 crc kubenswrapper[4690]: E1203 13:11:10.314795 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:11 crc kubenswrapper[4690]: I1203 13:11:11.314204 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:11 crc kubenswrapper[4690]: I1203 13:11:11.314262 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:11 crc kubenswrapper[4690]: E1203 13:11:11.314375 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:11 crc kubenswrapper[4690]: E1203 13:11:11.314466 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:12 crc kubenswrapper[4690]: I1203 13:11:12.315026 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:12 crc kubenswrapper[4690]: I1203 13:11:12.315046 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:12 crc kubenswrapper[4690]: E1203 13:11:12.315141 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:12 crc kubenswrapper[4690]: E1203 13:11:12.315250 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:13 crc kubenswrapper[4690]: I1203 13:11:13.313727 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:13 crc kubenswrapper[4690]: I1203 13:11:13.313759 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:13 crc kubenswrapper[4690]: E1203 13:11:13.314056 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:13 crc kubenswrapper[4690]: E1203 13:11:13.314239 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:14 crc kubenswrapper[4690]: E1203 13:11:14.262748 4690 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 13:11:14 crc kubenswrapper[4690]: I1203 13:11:14.313995 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:14 crc kubenswrapper[4690]: I1203 13:11:14.314046 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:14 crc kubenswrapper[4690]: E1203 13:11:14.315424 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:14 crc kubenswrapper[4690]: E1203 13:11:14.315578 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:14 crc kubenswrapper[4690]: E1203 13:11:14.415199 4690 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 13:11:15 crc kubenswrapper[4690]: I1203 13:11:15.314324 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:15 crc kubenswrapper[4690]: I1203 13:11:15.314406 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:15 crc kubenswrapper[4690]: E1203 13:11:15.314560 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:15 crc kubenswrapper[4690]: E1203 13:11:15.314628 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:16 crc kubenswrapper[4690]: I1203 13:11:16.314310 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:16 crc kubenswrapper[4690]: E1203 13:11:16.314505 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:16 crc kubenswrapper[4690]: I1203 13:11:16.314315 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:16 crc kubenswrapper[4690]: E1203 13:11:16.314705 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:17 crc kubenswrapper[4690]: I1203 13:11:17.314732 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:17 crc kubenswrapper[4690]: E1203 13:11:17.314900 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:17 crc kubenswrapper[4690]: I1203 13:11:17.315005 4690 scope.go:117] "RemoveContainer" containerID="f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6" Dec 03 13:11:17 crc kubenswrapper[4690]: I1203 13:11:17.315115 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:17 crc kubenswrapper[4690]: E1203 13:11:17.315296 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:17 crc kubenswrapper[4690]: I1203 13:11:17.342851 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-flm4q" podStartSLOduration=104.342831766 podStartE2EDuration="1m44.342831766s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:08.942235869 +0000 UTC m=+114.923156322" watchObservedRunningTime="2025-12-03 13:11:17.342831766 +0000 UTC m=+123.323752199" Dec 03 13:11:17 crc kubenswrapper[4690]: I1203 13:11:17.950554 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/1.log" Dec 03 13:11:17 crc kubenswrapper[4690]: I1203 13:11:17.951085 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tf2cb" event={"ID":"bf580211-e03a-444e-98b8-c12f99091b80","Type":"ContainerStarted","Data":"ffc489323eede1d26eeba3894b5d1ebc6e278d6fb487a0a0ce651b5a1d2f456f"} Dec 03 13:11:18 crc kubenswrapper[4690]: I1203 13:11:18.314242 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:18 crc kubenswrapper[4690]: E1203 13:11:18.314440 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:18 crc kubenswrapper[4690]: I1203 13:11:18.314277 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:18 crc kubenswrapper[4690]: E1203 13:11:18.314557 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:18 crc kubenswrapper[4690]: I1203 13:11:18.315357 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:11:18 crc kubenswrapper[4690]: I1203 13:11:18.957027 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/3.log" Dec 03 13:11:18 crc kubenswrapper[4690]: I1203 13:11:18.959519 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerStarted","Data":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} Dec 03 13:11:18 crc kubenswrapper[4690]: I1203 13:11:18.960008 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:11:18 crc kubenswrapper[4690]: I1203 13:11:18.995120 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podStartSLOduration=105.995101335 podStartE2EDuration="1m45.995101335s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:18.993703 +0000 UTC m=+124.974623433" watchObservedRunningTime="2025-12-03 13:11:18.995101335 +0000 UTC m=+124.976021768" Dec 03 13:11:19 crc kubenswrapper[4690]: I1203 13:11:19.064454 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mbbpr"] Dec 03 13:11:19 crc kubenswrapper[4690]: I1203 13:11:19.064621 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:19 crc kubenswrapper[4690]: E1203 13:11:19.064736 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:19 crc kubenswrapper[4690]: I1203 13:11:19.314148 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:19 crc kubenswrapper[4690]: E1203 13:11:19.314277 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:19 crc kubenswrapper[4690]: E1203 13:11:19.416699 4690 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 13:11:20 crc kubenswrapper[4690]: I1203 13:11:20.314056 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:20 crc kubenswrapper[4690]: I1203 13:11:20.314117 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:20 crc kubenswrapper[4690]: E1203 13:11:20.314414 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:20 crc kubenswrapper[4690]: E1203 13:11:20.314519 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:21 crc kubenswrapper[4690]: I1203 13:11:21.314622 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:21 crc kubenswrapper[4690]: I1203 13:11:21.314622 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:21 crc kubenswrapper[4690]: E1203 13:11:21.315329 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:21 crc kubenswrapper[4690]: E1203 13:11:21.315565 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:22 crc kubenswrapper[4690]: I1203 13:11:22.314717 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:22 crc kubenswrapper[4690]: E1203 13:11:22.314928 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:22 crc kubenswrapper[4690]: I1203 13:11:22.315053 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:22 crc kubenswrapper[4690]: E1203 13:11:22.315134 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:23 crc kubenswrapper[4690]: I1203 13:11:23.313703 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:23 crc kubenswrapper[4690]: I1203 13:11:23.313735 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:23 crc kubenswrapper[4690]: E1203 13:11:23.314176 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mbbpr" podUID="abab6c63-91d1-4f36-9be9-ae3b1a43742e" Dec 03 13:11:23 crc kubenswrapper[4690]: E1203 13:11:23.314213 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:11:24 crc kubenswrapper[4690]: I1203 13:11:24.314132 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:24 crc kubenswrapper[4690]: I1203 13:11:24.314236 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:24 crc kubenswrapper[4690]: E1203 13:11:24.314654 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:11:24 crc kubenswrapper[4690]: E1203 13:11:24.314781 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:11:25 crc kubenswrapper[4690]: I1203 13:11:25.314238 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:25 crc kubenswrapper[4690]: I1203 13:11:25.314312 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:25 crc kubenswrapper[4690]: I1203 13:11:25.316652 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 13:11:25 crc kubenswrapper[4690]: I1203 13:11:25.316759 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 13:11:25 crc kubenswrapper[4690]: I1203 13:11:25.317844 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 13:11:25 crc kubenswrapper[4690]: I1203 13:11:25.319313 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 13:11:26 crc kubenswrapper[4690]: I1203 13:11:26.314690 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:26 crc kubenswrapper[4690]: I1203 13:11:26.314993 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:26 crc kubenswrapper[4690]: I1203 13:11:26.317262 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 13:11:26 crc kubenswrapper[4690]: I1203 13:11:26.317309 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.619729 4690 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.682524 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kv48s"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.683343 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.685121 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.686032 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dsmtf"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.686203 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.686389 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.688240 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.689016 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.691969 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.692190 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.692368 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.692917 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.693178 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.693476 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.693823 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.694046 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.694240 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54f5g"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.694730 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.695163 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.695365 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.695537 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.696334 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.696593 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.697001 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.698337 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.699596 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w5nc4"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.700020 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.700552 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.700817 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.700924 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.701077 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.701225 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.701228 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.701301 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.701597 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.701740 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.701989 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.702108 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.702318 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.702422 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.702444 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.710438 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.710485 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.711112 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.711688 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.713068 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-xjjml"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.713713 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xjjml" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.714299 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.714845 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.715173 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.715923 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.716110 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.721327 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.721747 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.722625 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.724451 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.724629 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.724698 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.726360 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.727515 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.728436 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.729989 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.740671 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.741746 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.742040 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.743209 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.743327 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.743423 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.743616 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.743727 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.743820 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.743901 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.744240 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.744420 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.744564 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745047 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745422 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745527 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745613 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745719 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745803 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745813 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745945 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.745988 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.746147 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.746195 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.746155 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.746318 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.746356 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.746417 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.746469 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-m2l5h"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.747087 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.752955 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.754543 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.754608 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.755634 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.755777 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.755887 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.756016 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.757698 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.758100 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.758676 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.759289 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bl4h4"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.759711 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.761409 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.761465 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.761576 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.761659 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.761409 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.761799 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.761967 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.763089 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.764430 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.766052 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.766607 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-sqp4x"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.767292 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.767918 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.768027 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.768345 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.769295 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.771975 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.772683 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.773061 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.773318 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.818409 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.819405 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.820264 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jldld"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.824963 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.825281 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.826047 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.826211 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.826363 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.826481 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.826608 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.826737 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.826887 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.827085 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.827149 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.827761 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.833000 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.833302 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.833442 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.833720 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.833798 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.833890 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.834178 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.834922 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.835220 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.835835 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.836259 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kv48s"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.836282 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9clnc"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.836628 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.836829 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.838244 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.839402 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.840204 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.840409 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.840965 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.840976 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.841244 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.841265 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.841372 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.841665 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.843484 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lds4l"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.844150 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-b2d5g"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.844764 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.844842 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.847302 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.848748 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.849065 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4pql2"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.849831 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.851424 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4mtxh"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.852179 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.852550 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854677 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dcea818d-07a3-4c60-8571-c71af537ca5c-audit-dir\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854736 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/32821680-989c-49f5-9b31-b5efed94d3c3-machine-approver-tls\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854771 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854791 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32821680-989c-49f5-9b31-b5efed94d3c3-auth-proxy-config\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854814 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3458d878-4d3a-4989-abe4-22770cec182e-proxy-tls\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854836 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/349d5d37-327a-48e7-bf3d-51e0cfb79c31-apiservice-cert\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854858 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwmms\" (UniqueName: \"kubernetes.io/projected/3fbba057-1569-44f3-9eb7-1ce074ae625e-kube-api-access-mwmms\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854898 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dcea818d-07a3-4c60-8571-c71af537ca5c-node-pullsecrets\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854918 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvz58\" (UniqueName: \"kubernetes.io/projected/349d5d37-327a-48e7-bf3d-51e0cfb79c31-kube-api-access-nvz58\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854942 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854965 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-config\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.854986 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtzpq\" (UniqueName: \"kubernetes.io/projected/32821680-989c-49f5-9b31-b5efed94d3c3-kube-api-access-jtzpq\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855007 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-policies\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855026 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-config\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855047 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855072 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjdll\" (UniqueName: \"kubernetes.io/projected/1778fa3f-41b1-4a29-a253-6b0be39e11c6-kube-api-access-vjdll\") pod \"cluster-samples-operator-665b6dd947-bcqd7\" (UID: \"1778fa3f-41b1-4a29-a253-6b0be39e11c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855093 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855113 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4xxl\" (UniqueName: \"kubernetes.io/projected/d5c2b268-0de1-4c88-b439-5292c56044dd-kube-api-access-x4xxl\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855144 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71232879-c2e3-4d23-b663-5b3cb4bf1d31-config\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855165 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg82r\" (UniqueName: \"kubernetes.io/projected/71232879-c2e3-4d23-b663-5b3cb4bf1d31-kube-api-access-dg82r\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855188 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855207 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32821680-989c-49f5-9b31-b5efed94d3c3-config\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855243 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71232879-c2e3-4d23-b663-5b3cb4bf1d31-trusted-ca\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855267 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855287 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-etcd-client\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855311 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsr5v\" (UniqueName: \"kubernetes.io/projected/75d357bf-d7e7-43cf-904b-0589a663f831-kube-api-access-gsr5v\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855336 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3fbba057-1569-44f3-9eb7-1ce074ae625e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855374 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f566aea3-313b-4311-ab29-491756446da4-signing-cabundle\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855396 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75d357bf-d7e7-43cf-904b-0589a663f831-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855420 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3458d878-4d3a-4989-abe4-22770cec182e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855447 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855484 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-encryption-config\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855507 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-serving-cert\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855529 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e35d957c-079a-484a-9508-2a1def9b682d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855549 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e35d957c-079a-484a-9508-2a1def9b682d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855572 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855594 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/60890c93-5ee6-4b32-af02-acc98b37dbd3-profile-collector-cert\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855614 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75d357bf-d7e7-43cf-904b-0589a663f831-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855631 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855648 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e35d957c-079a-484a-9508-2a1def9b682d-config\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855666 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1778fa3f-41b1-4a29-a253-6b0be39e11c6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bcqd7\" (UID: \"1778fa3f-41b1-4a29-a253-6b0be39e11c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855683 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/60890c93-5ee6-4b32-af02-acc98b37dbd3-srv-cert\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855700 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fbba057-1569-44f3-9eb7-1ce074ae625e-serving-cert\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855716 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0922d35f-705f-48af-823a-15155eecabc9-srv-cert\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855731 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqqmv\" (UniqueName: \"kubernetes.io/projected/3458d878-4d3a-4989-abe4-22770cec182e-kube-api-access-wqqmv\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855753 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-dir\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855770 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855790 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2gz7\" (UniqueName: \"kubernetes.io/projected/f566aea3-313b-4311-ab29-491756446da4-kube-api-access-f2gz7\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855805 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-audit\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855824 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71232879-c2e3-4d23-b663-5b3cb4bf1d31-serving-cert\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855840 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h64s7\" (UniqueName: \"kubernetes.io/projected/0922d35f-705f-48af-823a-15155eecabc9-kube-api-access-h64s7\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.855894 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/349d5d37-327a-48e7-bf3d-51e0cfb79c31-tmpfs\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856054 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/349d5d37-327a-48e7-bf3d-51e0cfb79c31-webhook-cert\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856113 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vwgw\" (UniqueName: \"kubernetes.io/projected/2b122759-56c3-4ee4-81dd-cf055ced8121-kube-api-access-8vwgw\") pod \"dns-operator-744455d44c-w5nc4\" (UID: \"2b122759-56c3-4ee4-81dd-cf055ced8121\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856139 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856170 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0922d35f-705f-48af-823a-15155eecabc9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856219 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dn9k\" (UniqueName: \"kubernetes.io/projected/60890c93-5ee6-4b32-af02-acc98b37dbd3-kube-api-access-6dn9k\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856250 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f566aea3-313b-4311-ab29-491756446da4-signing-key\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856269 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856323 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856349 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6gg9\" (UniqueName: \"kubernetes.io/projected/dcea818d-07a3-4c60-8571-c71af537ca5c-kube-api-access-r6gg9\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856382 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-image-import-ca\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856401 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856408 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856426 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856462 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbbdt\" (UniqueName: \"kubernetes.io/projected/cd786c95-6f85-4a6e-a7c7-341666f2b478-kube-api-access-tbbdt\") pod \"downloads-7954f5f757-xjjml\" (UID: \"cd786c95-6f85-4a6e-a7c7-341666f2b478\") " pod="openshift-console/downloads-7954f5f757-xjjml" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856478 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b122759-56c3-4ee4-81dd-cf055ced8121-metrics-tls\") pod \"dns-operator-744455d44c-w5nc4\" (UID: \"2b122759-56c3-4ee4-81dd-cf055ced8121\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.856576 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.857299 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.858035 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bhp9h"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.858488 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.861027 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jdlts"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.863335 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.866963 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.867692 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.867720 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54f5g"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.867815 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.869835 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dsmtf"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.870948 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.871648 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.872548 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w5nc4"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.876178 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.877588 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.879186 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xjjml"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.879877 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.881001 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-42cnp"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.882221 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.883404 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.885893 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.886106 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.887749 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.888779 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-b2d5g"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.890831 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sqp4x"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.892133 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.893138 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.894341 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.896157 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.897321 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-m2l5h"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.898605 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lds4l"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.902049 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.903033 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.904531 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-42cnp"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.905909 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.907073 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.911673 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.914692 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.920239 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.920330 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.922219 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4mtxh"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.924366 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.927111 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.927738 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bhp9h"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.928756 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-ldbtd"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.929987 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jdlts"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.930108 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ldbtd" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.931702 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jldld"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.932948 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-vhrgj"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.934297 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9clnc"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.934394 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.935360 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4pql2"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.936691 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ldbtd"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.937904 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.944803 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-46m2s"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.945903 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.950146 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.959792 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e35d957c-079a-484a-9508-2a1def9b682d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.959839 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e35d957c-079a-484a-9508-2a1def9b682d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.959885 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.959916 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/281faa96-ab8c-4fa4-886c-b69f1d818c9e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.959943 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/60890c93-5ee6-4b32-af02-acc98b37dbd3-profile-collector-cert\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.959965 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75d357bf-d7e7-43cf-904b-0589a663f831-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.959987 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960009 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0922d35f-705f-48af-823a-15155eecabc9-srv-cert\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960030 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqqmv\" (UniqueName: \"kubernetes.io/projected/3458d878-4d3a-4989-abe4-22770cec182e-kube-api-access-wqqmv\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960051 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e35d957c-079a-484a-9508-2a1def9b682d-config\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960075 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1778fa3f-41b1-4a29-a253-6b0be39e11c6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bcqd7\" (UID: \"1778fa3f-41b1-4a29-a253-6b0be39e11c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960133 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/60890c93-5ee6-4b32-af02-acc98b37dbd3-srv-cert\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960168 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fbba057-1569-44f3-9eb7-1ce074ae625e-serving-cert\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960204 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-dir\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960221 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960247 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a2df869-a97d-4371-ae6c-fa30342f98df-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960269 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960288 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a2df869-a97d-4371-ae6c-fa30342f98df-images\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960304 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-audit\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960322 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71232879-c2e3-4d23-b663-5b3cb4bf1d31-serving-cert\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960341 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2gz7\" (UniqueName: \"kubernetes.io/projected/f566aea3-313b-4311-ab29-491756446da4-kube-api-access-f2gz7\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960358 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vwgw\" (UniqueName: \"kubernetes.io/projected/2b122759-56c3-4ee4-81dd-cf055ced8121-kube-api-access-8vwgw\") pod \"dns-operator-744455d44c-w5nc4\" (UID: \"2b122759-56c3-4ee4-81dd-cf055ced8121\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960374 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h64s7\" (UniqueName: \"kubernetes.io/projected/0922d35f-705f-48af-823a-15155eecabc9-kube-api-access-h64s7\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960390 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/349d5d37-327a-48e7-bf3d-51e0cfb79c31-tmpfs\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960407 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/349d5d37-327a-48e7-bf3d-51e0cfb79c31-webhook-cert\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960425 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960442 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0922d35f-705f-48af-823a-15155eecabc9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960460 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dn9k\" (UniqueName: \"kubernetes.io/projected/60890c93-5ee6-4b32-af02-acc98b37dbd3-kube-api-access-6dn9k\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960483 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f566aea3-313b-4311-ab29-491756446da4-signing-key\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960500 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960520 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960541 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-image-import-ca\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960557 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6gg9\" (UniqueName: \"kubernetes.io/projected/dcea818d-07a3-4c60-8571-c71af537ca5c-kube-api-access-r6gg9\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960578 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrjp8\" (UniqueName: \"kubernetes.io/projected/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-kube-api-access-hrjp8\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960603 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960623 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960640 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbbdt\" (UniqueName: \"kubernetes.io/projected/cd786c95-6f85-4a6e-a7c7-341666f2b478-kube-api-access-tbbdt\") pod \"downloads-7954f5f757-xjjml\" (UID: \"cd786c95-6f85-4a6e-a7c7-341666f2b478\") " pod="openshift-console/downloads-7954f5f757-xjjml" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960656 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b122759-56c3-4ee4-81dd-cf055ced8121-metrics-tls\") pod \"dns-operator-744455d44c-w5nc4\" (UID: \"2b122759-56c3-4ee4-81dd-cf055ced8121\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960681 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dcea818d-07a3-4c60-8571-c71af537ca5c-audit-dir\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960699 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/32821680-989c-49f5-9b31-b5efed94d3c3-machine-approver-tls\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960717 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960736 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32821680-989c-49f5-9b31-b5efed94d3c3-auth-proxy-config\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960751 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3458d878-4d3a-4989-abe4-22770cec182e-proxy-tls\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960770 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/349d5d37-327a-48e7-bf3d-51e0cfb79c31-apiservice-cert\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960785 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960803 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960819 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960836 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwmms\" (UniqueName: \"kubernetes.io/projected/3fbba057-1569-44f3-9eb7-1ce074ae625e-kube-api-access-mwmms\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960856 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1a2df869-a97d-4371-ae6c-fa30342f98df-proxy-tls\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960880 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960899 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dcea818d-07a3-4c60-8571-c71af537ca5c-node-pullsecrets\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960918 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvz58\" (UniqueName: \"kubernetes.io/projected/349d5d37-327a-48e7-bf3d-51e0cfb79c31-kube-api-access-nvz58\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960934 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtzpq\" (UniqueName: \"kubernetes.io/projected/32821680-989c-49f5-9b31-b5efed94d3c3-kube-api-access-jtzpq\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960964 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960987 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-config\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961011 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-config\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961026 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-policies\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961042 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdrvm\" (UniqueName: \"kubernetes.io/projected/1a2df869-a97d-4371-ae6c-fa30342f98df-kube-api-access-bdrvm\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961060 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961077 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjdll\" (UniqueName: \"kubernetes.io/projected/1778fa3f-41b1-4a29-a253-6b0be39e11c6-kube-api-access-vjdll\") pod \"cluster-samples-operator-665b6dd947-bcqd7\" (UID: \"1778fa3f-41b1-4a29-a253-6b0be39e11c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961092 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4xxl\" (UniqueName: \"kubernetes.io/projected/d5c2b268-0de1-4c88-b439-5292c56044dd-kube-api-access-x4xxl\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961112 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961137 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71232879-c2e3-4d23-b663-5b3cb4bf1d31-config\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961159 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg82r\" (UniqueName: \"kubernetes.io/projected/71232879-c2e3-4d23-b663-5b3cb4bf1d31-kube-api-access-dg82r\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961185 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961211 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32821680-989c-49f5-9b31-b5efed94d3c3-config\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961230 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71232879-c2e3-4d23-b663-5b3cb4bf1d31-trusted-ca\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961252 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961268 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-etcd-client\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961286 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsr5v\" (UniqueName: \"kubernetes.io/projected/75d357bf-d7e7-43cf-904b-0589a663f831-kube-api-access-gsr5v\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961305 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/281faa96-ab8c-4fa4-886c-b69f1d818c9e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961327 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3fbba057-1569-44f3-9eb7-1ce074ae625e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961366 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961394 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f566aea3-313b-4311-ab29-491756446da4-signing-cabundle\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961412 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75d357bf-d7e7-43cf-904b-0589a663f831-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961432 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3458d878-4d3a-4989-abe4-22770cec182e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961452 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961471 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961485 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961493 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e35d957c-079a-484a-9508-2a1def9b682d-config\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961503 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpxf7\" (UniqueName: \"kubernetes.io/projected/281faa96-ab8c-4fa4-886c-b69f1d818c9e-kube-api-access-fpxf7\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961523 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-encryption-config\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.961544 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlbxz\" (UniqueName: \"kubernetes.io/projected/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-kube-api-access-zlbxz\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.963699 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75d357bf-d7e7-43cf-904b-0589a663f831-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.964117 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3fbba057-1569-44f3-9eb7-1ce074ae625e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.964787 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f566aea3-313b-4311-ab29-491756446da4-signing-cabundle\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.966036 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.966367 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.966498 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-serving-cert\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.966843 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71232879-c2e3-4d23-b663-5b3cb4bf1d31-config\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.967260 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-audit\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.967399 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-encryption-config\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.967509 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b122759-56c3-4ee4-81dd-cf055ced8121-metrics-tls\") pod \"dns-operator-744455d44c-w5nc4\" (UID: \"2b122759-56c3-4ee4-81dd-cf055ced8121\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.967844 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75d357bf-d7e7-43cf-904b-0589a663f831-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.968022 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3458d878-4d3a-4989-abe4-22770cec182e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.968264 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/349d5d37-327a-48e7-bf3d-51e0cfb79c31-tmpfs\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.968342 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71232879-c2e3-4d23-b663-5b3cb4bf1d31-trusted-ca\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.968677 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32821680-989c-49f5-9b31-b5efed94d3c3-config\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.968683 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.968958 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dcea818d-07a3-4c60-8571-c71af537ca5c-audit-dir\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.969311 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-image-import-ca\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.960501 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-46m2s"] Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.969439 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-dir\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.969485 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dcea818d-07a3-4c60-8571-c71af537ca5c-node-pullsecrets\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.970136 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.970360 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0922d35f-705f-48af-823a-15155eecabc9-srv-cert\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.970795 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32821680-989c-49f5-9b31-b5efed94d3c3-auth-proxy-config\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.971236 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.971369 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.971662 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-policies\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.971662 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1778fa3f-41b1-4a29-a253-6b0be39e11c6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bcqd7\" (UID: \"1778fa3f-41b1-4a29-a253-6b0be39e11c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.971854 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-config\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.971996 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-config\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.973080 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/60890c93-5ee6-4b32-af02-acc98b37dbd3-profile-collector-cert\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.973524 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.973588 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f566aea3-313b-4311-ab29-491756446da4-signing-key\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.973655 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-serving-cert\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.973842 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcea818d-07a3-4c60-8571-c71af537ca5c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.973977 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.974200 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/60890c93-5ee6-4b32-af02-acc98b37dbd3-srv-cert\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.974240 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/349d5d37-327a-48e7-bf3d-51e0cfb79c31-webhook-cert\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.974424 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71232879-c2e3-4d23-b663-5b3cb4bf1d31-serving-cert\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.974507 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.974832 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/32821680-989c-49f5-9b31-b5efed94d3c3-machine-approver-tls\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.975082 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dcea818d-07a3-4c60-8571-c71af537ca5c-etcd-client\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.975456 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0922d35f-705f-48af-823a-15155eecabc9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.975453 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/349d5d37-327a-48e7-bf3d-51e0cfb79c31-apiservice-cert\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.975913 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e35d957c-079a-484a-9508-2a1def9b682d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.975961 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.975986 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3458d878-4d3a-4989-abe4-22770cec182e-proxy-tls\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.976813 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.976848 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.977356 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fbba057-1569-44f3-9eb7-1ce074ae625e-serving-cert\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.983167 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:28 crc kubenswrapper[4690]: I1203 13:11:28.985841 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.006080 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.026304 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.045572 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.066039 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067491 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/281faa96-ab8c-4fa4-886c-b69f1d818c9e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067531 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067571 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067595 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067622 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpxf7\" (UniqueName: \"kubernetes.io/projected/281faa96-ab8c-4fa4-886c-b69f1d818c9e-kube-api-access-fpxf7\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067646 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlbxz\" (UniqueName: \"kubernetes.io/projected/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-kube-api-access-zlbxz\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067683 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/281faa96-ab8c-4fa4-886c-b69f1d818c9e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067727 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a2df869-a97d-4371-ae6c-fa30342f98df-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067748 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067767 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a2df869-a97d-4371-ae6c-fa30342f98df-images\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067848 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrjp8\" (UniqueName: \"kubernetes.io/projected/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-kube-api-access-hrjp8\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067929 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067965 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.067991 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.068031 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1a2df869-a97d-4371-ae6c-fa30342f98df-proxy-tls\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.068080 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdrvm\" (UniqueName: \"kubernetes.io/projected/1a2df869-a97d-4371-ae6c-fa30342f98df-kube-api-access-bdrvm\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.068916 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/281faa96-ab8c-4fa4-886c-b69f1d818c9e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.070741 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a2df869-a97d-4371-ae6c-fa30342f98df-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.072272 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/281faa96-ab8c-4fa4-886c-b69f1d818c9e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.085805 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.106339 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.125668 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.145680 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.171837 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.179988 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.209051 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.213233 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.225663 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.246968 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.266655 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.287021 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.307283 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.326449 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.347423 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.366825 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.386285 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.407201 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.426312 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.446452 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.466663 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.470777 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a2df869-a97d-4371-ae6c-fa30342f98df-images\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.486835 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.506377 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.526597 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.546258 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.554591 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1a2df869-a97d-4371-ae6c-fa30342f98df-proxy-tls\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.566853 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.592478 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.606671 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.626476 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.646572 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.666122 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.685495 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.705719 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.725275 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.745855 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.765766 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.785937 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.806770 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.826719 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.846586 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.864674 4690 request.go:700] Waited for 1.014497675s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.866255 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.886284 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.906977 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.926762 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.945906 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.966201 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 13:11:29 crc kubenswrapper[4690]: I1203 13:11:29.993131 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.005793 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.026596 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.046135 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.067635 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: E1203 13:11:30.069623 4690 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 03 13:11:30 crc kubenswrapper[4690]: E1203 13:11:30.069637 4690 secret.go:188] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 13:11:30 crc kubenswrapper[4690]: E1203 13:11:30.069675 4690 configmap.go:193] Couldn't get configMap openshift-controller-manager/openshift-global-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 13:11:30 crc kubenswrapper[4690]: E1203 13:11:30.069703 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config podName:4975e4cf-5be0-40e1-a4b0-2db6ec43c207 nodeName:}" failed. No retries permitted until 2025-12-03 13:11:30.569677122 +0000 UTC m=+136.550597555 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config") pod "controller-manager-879f6c89f-jdlts" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207") : failed to sync configmap cache: timed out waiting for the condition Dec 03 13:11:30 crc kubenswrapper[4690]: E1203 13:11:30.069715 4690 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 13:11:30 crc kubenswrapper[4690]: E1203 13:11:30.069725 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles podName:4975e4cf-5be0-40e1-a4b0-2db6ec43c207 nodeName:}" failed. No retries permitted until 2025-12-03 13:11:30.569712863 +0000 UTC m=+136.550633296 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-ca-bundles" (UniqueName: "kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles") pod "controller-manager-879f6c89f-jdlts" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207") : failed to sync configmap cache: timed out waiting for the condition Dec 03 13:11:30 crc kubenswrapper[4690]: E1203 13:11:30.069747 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca podName:4975e4cf-5be0-40e1-a4b0-2db6ec43c207 nodeName:}" failed. No retries permitted until 2025-12-03 13:11:30.569732314 +0000 UTC m=+136.550652747 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca") pod "controller-manager-879f6c89f-jdlts" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207") : failed to sync configmap cache: timed out waiting for the condition Dec 03 13:11:30 crc kubenswrapper[4690]: E1203 13:11:30.069764 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert podName:4975e4cf-5be0-40e1-a4b0-2db6ec43c207 nodeName:}" failed. No retries permitted until 2025-12-03 13:11:30.569756154 +0000 UTC m=+136.550676587 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert") pod "controller-manager-879f6c89f-jdlts" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207") : failed to sync secret cache: timed out waiting for the condition Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.086214 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.107421 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.125375 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.145669 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.165949 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.186436 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.206391 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.226068 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.253248 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.266022 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.286350 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.306825 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.326826 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.345577 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.372478 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.387412 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.406473 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.426796 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.447365 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.487149 4690 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.506103 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.526859 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.546587 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.567378 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.593997 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.594584 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.594916 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.595143 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.596102 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.596743 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.596749 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.597986 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.599321 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.606163 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.626826 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.647364 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.666552 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.685392 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.705622 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.725288 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.760664 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e35d957c-079a-484a-9508-2a1def9b682d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hzthl\" (UID: \"e35d957c-079a-484a-9508-2a1def9b682d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.783391 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqqmv\" (UniqueName: \"kubernetes.io/projected/3458d878-4d3a-4989-abe4-22770cec182e-kube-api-access-wqqmv\") pod \"machine-config-controller-84d6567774-jwh9v\" (UID: \"3458d878-4d3a-4989-abe4-22770cec182e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.801484 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsr5v\" (UniqueName: \"kubernetes.io/projected/75d357bf-d7e7-43cf-904b-0589a663f831-kube-api-access-gsr5v\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4bjl\" (UID: \"75d357bf-d7e7-43cf-904b-0589a663f831\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.822005 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjdll\" (UniqueName: \"kubernetes.io/projected/1778fa3f-41b1-4a29-a253-6b0be39e11c6-kube-api-access-vjdll\") pod \"cluster-samples-operator-665b6dd947-bcqd7\" (UID: \"1778fa3f-41b1-4a29-a253-6b0be39e11c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.843690 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4xxl\" (UniqueName: \"kubernetes.io/projected/d5c2b268-0de1-4c88-b439-5292c56044dd-kube-api-access-x4xxl\") pod \"oauth-openshift-558db77b4-54f5g\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.860669 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2gz7\" (UniqueName: \"kubernetes.io/projected/f566aea3-313b-4311-ab29-491756446da4-kube-api-access-f2gz7\") pod \"service-ca-9c57cc56f-m2l5h\" (UID: \"f566aea3-313b-4311-ab29-491756446da4\") " pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.880951 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg82r\" (UniqueName: \"kubernetes.io/projected/71232879-c2e3-4d23-b663-5b3cb4bf1d31-kube-api-access-dg82r\") pod \"console-operator-58897d9998-dsmtf\" (UID: \"71232879-c2e3-4d23-b663-5b3cb4bf1d31\") " pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.884535 4690 request.go:700] Waited for 1.917461893s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/serviceaccounts/olm-operator-serviceaccount/token Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.901126 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h64s7\" (UniqueName: \"kubernetes.io/projected/0922d35f-705f-48af-823a-15155eecabc9-kube-api-access-h64s7\") pod \"olm-operator-6b444d44fb-98mfv\" (UID: \"0922d35f-705f-48af-823a-15155eecabc9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.920155 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vwgw\" (UniqueName: \"kubernetes.io/projected/2b122759-56c3-4ee4-81dd-cf055ced8121-kube-api-access-8vwgw\") pod \"dns-operator-744455d44c-w5nc4\" (UID: \"2b122759-56c3-4ee4-81dd-cf055ced8121\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.938464 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.941084 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6gg9\" (UniqueName: \"kubernetes.io/projected/dcea818d-07a3-4c60-8571-c71af537ca5c-kube-api-access-r6gg9\") pod \"apiserver-76f77b778f-kv48s\" (UID: \"dcea818d-07a3-4c60-8571-c71af537ca5c\") " pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.942979 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.964843 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtzpq\" (UniqueName: \"kubernetes.io/projected/32821680-989c-49f5-9b31-b5efed94d3c3-kube-api-access-jtzpq\") pod \"machine-approver-56656f9798-w76dr\" (UID: \"32821680-989c-49f5-9b31-b5efed94d3c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.976783 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.981127 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwmms\" (UniqueName: \"kubernetes.io/projected/3fbba057-1569-44f3-9eb7-1ce074ae625e-kube-api-access-mwmms\") pod \"openshift-config-operator-7777fb866f-z8cfl\" (UID: \"3fbba057-1569-44f3-9eb7-1ce074ae625e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:30 crc kubenswrapper[4690]: I1203 13:11:30.985250 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.002216 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.003226 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvz58\" (UniqueName: \"kubernetes.io/projected/349d5d37-327a-48e7-bf3d-51e0cfb79c31-kube-api-access-nvz58\") pod \"packageserver-d55dfcdfc-9c5gx\" (UID: \"349d5d37-327a-48e7-bf3d-51e0cfb79c31\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.010748 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.026125 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbbdt\" (UniqueName: \"kubernetes.io/projected/cd786c95-6f85-4a6e-a7c7-341666f2b478-kube-api-access-tbbdt\") pod \"downloads-7954f5f757-xjjml\" (UID: \"cd786c95-6f85-4a6e-a7c7-341666f2b478\") " pod="openshift-console/downloads-7954f5f757-xjjml" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.043343 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dn9k\" (UniqueName: \"kubernetes.io/projected/60890c93-5ee6-4b32-af02-acc98b37dbd3-kube-api-access-6dn9k\") pod \"catalog-operator-68c6474976-69qxx\" (UID: \"60890c93-5ee6-4b32-af02-acc98b37dbd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.052073 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.061006 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.062175 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/907aeacd-adcc-4d58-b59e-e4d1c45f56e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lsjpb\" (UID: \"907aeacd-adcc-4d58-b59e-e4d1c45f56e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.086534 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdrvm\" (UniqueName: \"kubernetes.io/projected/1a2df869-a97d-4371-ae6c-fa30342f98df-kube-api-access-bdrvm\") pod \"machine-config-operator-74547568cd-7q4tf\" (UID: \"1a2df869-a97d-4371-ae6c-fa30342f98df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.102000 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.103740 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpxf7\" (UniqueName: \"kubernetes.io/projected/281faa96-ab8c-4fa4-886c-b69f1d818c9e-kube-api-access-fpxf7\") pod \"openshift-apiserver-operator-796bbdcf4f-brjbq\" (UID: \"281faa96-ab8c-4fa4-886c-b69f1d818c9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.117953 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.123933 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlbxz\" (UniqueName: \"kubernetes.io/projected/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-kube-api-access-zlbxz\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.148579 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.149218 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.150499 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrjp8\" (UniqueName: \"kubernetes.io/projected/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-kube-api-access-hrjp8\") pod \"controller-manager-879f6c89f-jdlts\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.175635 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.176722 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d8a29fb-3ded-4198-8da3-dd09fc8bfad5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h4snn\" (UID: \"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.185372 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54f5g"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.198796 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.201825 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tl2k\" (UniqueName: \"kubernetes.io/projected/bf9b1cbd-2048-4999-860c-5fab190dd3fb-kube-api-access-8tl2k\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.201919 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-bound-sa-token\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.201944 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.201970 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6wtw\" (UniqueName: \"kubernetes.io/projected/61a7f75a-3158-4e5f-a5f9-5c123079888a-kube-api-access-x6wtw\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.201989 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-service-ca-bundle\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202007 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-metrics-certs\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202030 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-certificates\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202090 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c499a600-3051-43fd-b1e8-24bc9f84f38f-serving-cert\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202114 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202138 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c499a600-3051-43fd-b1e8-24bc9f84f38f-config\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202164 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-client\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202222 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj9fm\" (UniqueName: \"kubernetes.io/projected/1c0c0215-8cde-4704-9ece-1d458f1efd49-kube-api-access-kj9fm\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202248 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-ca\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202295 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdbx9\" (UniqueName: \"kubernetes.io/projected/da998e2c-9c36-4f73-a839-d81ad59c955d-kube-api-access-mdbx9\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202320 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202353 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c0c0215-8cde-4704-9ece-1d458f1efd49-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202406 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-serving-cert\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202430 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nds6\" (UniqueName: \"kubernetes.io/projected/c499a600-3051-43fd-b1e8-24bc9f84f38f-kube-api-access-8nds6\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202465 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tkw6\" (UniqueName: \"kubernetes.io/projected/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-kube-api-access-8tkw6\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202505 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/23e81a05-bdfd-4a30-82b0-9eded242e91c-secret-volume\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202529 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-default-certificate\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.202566 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-oauth-serving-cert\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.203754 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-encryption-config\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.204062 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/23e81a05-bdfd-4a30-82b0-9eded242e91c-config-volume\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.204116 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-serving-cert\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.204147 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.204182 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-689m7\" (UniqueName: \"kubernetes.io/projected/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-kube-api-access-689m7\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.204679 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.204721 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-images\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.204808 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e4b586d8-5bf2-482b-8721-e8385575b53a-audit-dir\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205003 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61a7f75a-3158-4e5f-a5f9-5c123079888a-trusted-ca\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205057 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-tls\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205076 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-trusted-ca-bundle\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205097 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckfpz\" (UniqueName: \"kubernetes.io/projected/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-kube-api-access-ckfpz\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205123 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-stats-auth\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205335 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-etcd-client\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205360 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-config\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205391 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxf9g\" (UniqueName: \"kubernetes.io/projected/37ddf5df-f54d-47f9-8aac-04a2bc93bc25-kube-api-access-kxf9g\") pod \"migrator-59844c95c7-sz52n\" (UID: \"37ddf5df-f54d-47f9-8aac-04a2bc93bc25\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205414 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205431 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-audit-policies\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205447 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-serving-cert\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205463 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqz2v\" (UniqueName: \"kubernetes.io/projected/e4b586d8-5bf2-482b-8721-e8385575b53a-kube-api-access-jqz2v\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205515 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfqq7\" (UniqueName: \"kubernetes.io/projected/4b08791c-eff7-4b2b-8d98-f37e0bdbe887-kube-api-access-nfqq7\") pod \"multus-admission-controller-857f4d67dd-b2d5g\" (UID: \"4b08791c-eff7-4b2b-8d98-f37e0bdbe887\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205551 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205572 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-service-ca\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205587 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.205611 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-client-ca\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.210233 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c077bf27-95bf-4abb-b285-8a62e2118c65-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jlxcw\" (UID: \"c077bf27-95bf-4abb-b285-8a62e2118c65\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.210314 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4b08791c-eff7-4b2b-8d98-f37e0bdbe887-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-b2d5g\" (UID: \"4b08791c-eff7-4b2b-8d98-f37e0bdbe887\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.210571 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.210702 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.211597 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.211646 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c0c0215-8cde-4704-9ece-1d458f1efd49-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.212102 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jvcdk\" (UID: \"2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.212211 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da998e2c-9c36-4f73-a839-d81ad59c955d-serving-cert\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.212266 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-service-ca\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.213175 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61a7f75a-3158-4e5f-a5f9-5c123079888a-metrics-tls\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.213265 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.213344 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-config\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.213378 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxfs6\" (UniqueName: \"kubernetes.io/projected/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-kube-api-access-dxfs6\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.213583 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pb56\" (UniqueName: \"kubernetes.io/projected/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-kube-api-access-9pb56\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.213857 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-oauth-config\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.213960 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6lvq\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-kube-api-access-t6lvq\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.213990 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rkn8\" (UniqueName: \"kubernetes.io/projected/c077bf27-95bf-4abb-b285-8a62e2118c65-kube-api-access-6rkn8\") pod \"package-server-manager-789f6589d5-jlxcw\" (UID: \"c077bf27-95bf-4abb-b285-8a62e2118c65\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.214063 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-config\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.214611 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8ltq\" (UniqueName: \"kubernetes.io/projected/2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3-kube-api-access-f8ltq\") pod \"control-plane-machine-set-operator-78cbb6b69f-jvcdk\" (UID: \"2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.214711 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-config\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.214922 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.215013 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf9b1cbd-2048-4999-860c-5fab190dd3fb-serving-cert\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.215289 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:31.715270931 +0000 UTC m=+137.696191364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.215650 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.215690 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fpgm\" (UniqueName: \"kubernetes.io/projected/23e81a05-bdfd-4a30-82b0-9eded242e91c-kube-api-access-2fpgm\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.215714 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61a7f75a-3158-4e5f-a5f9-5c123079888a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.218700 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-trusted-ca\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.218842 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-config\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: W1203 13:11:31.230881 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5c2b268_0de1_4c88_b439_5292c56044dd.slice/crio-70f4c4da7dce26de4d5962dce8f7781ffeecb4ddbef91635d525650ef02f0e21 WatchSource:0}: Error finding container 70f4c4da7dce26de4d5962dce8f7781ffeecb4ddbef91635d525650ef02f0e21: Status 404 returned error can't find the container with id 70f4c4da7dce26de4d5962dce8f7781ffeecb4ddbef91635d525650ef02f0e21 Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.261389 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.275554 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.294340 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xjjml" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.320260 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.320932 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.320965 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c499a600-3051-43fd-b1e8-24bc9f84f38f-config\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.320994 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvbkn\" (UniqueName: \"kubernetes.io/projected/2d188da8-708a-4cdb-a311-39896852177d-kube-api-access-cvbkn\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321019 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-client\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321047 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d761d3d-9962-4334-916a-dcaf73bfc51f-cert\") pod \"ingress-canary-ldbtd\" (UID: \"4d761d3d-9962-4334-916a-dcaf73bfc51f\") " pod="openshift-ingress-canary/ingress-canary-ldbtd" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321070 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj9fm\" (UniqueName: \"kubernetes.io/projected/1c0c0215-8cde-4704-9ece-1d458f1efd49-kube-api-access-kj9fm\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321094 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-ca\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321124 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdbx9\" (UniqueName: \"kubernetes.io/projected/da998e2c-9c36-4f73-a839-d81ad59c955d-kube-api-access-mdbx9\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321143 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321174 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c0c0215-8cde-4704-9ece-1d458f1efd49-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321201 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-serving-cert\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321227 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nds6\" (UniqueName: \"kubernetes.io/projected/c499a600-3051-43fd-b1e8-24bc9f84f38f-kube-api-access-8nds6\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321297 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tkw6\" (UniqueName: \"kubernetes.io/projected/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-kube-api-access-8tkw6\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321323 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/23e81a05-bdfd-4a30-82b0-9eded242e91c-secret-volume\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321349 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kczns\" (UniqueName: \"kubernetes.io/projected/4d761d3d-9962-4334-916a-dcaf73bfc51f-kube-api-access-kczns\") pod \"ingress-canary-ldbtd\" (UID: \"4d761d3d-9962-4334-916a-dcaf73bfc51f\") " pod="openshift-ingress-canary/ingress-canary-ldbtd" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321379 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-default-certificate\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321400 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/16dd8be0-522a-4cbc-b837-0562192609b9-certs\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321423 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-socket-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321447 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-oauth-serving-cert\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321471 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-encryption-config\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321497 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-plugins-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321519 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-registration-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.321790 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-mountpoint-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.323441 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.326685 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/23e81a05-bdfd-4a30-82b0-9eded242e91c-config-volume\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.326785 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-689m7\" (UniqueName: \"kubernetes.io/projected/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-kube-api-access-689m7\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.326808 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/16dd8be0-522a-4cbc-b837-0562192609b9-node-bootstrap-token\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.326836 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-serving-cert\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.326853 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.326981 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-images\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.328090 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-ca\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.328454 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-oauth-serving-cert\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.328486 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/23e81a05-bdfd-4a30-82b0-9eded242e91c-config-volume\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.328714 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.328773 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d188da8-708a-4cdb-a311-39896852177d-config-volume\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.328786 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c499a600-3051-43fd-b1e8-24bc9f84f38f-config\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.328822 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e4b586d8-5bf2-482b-8721-e8385575b53a-audit-dir\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.329113 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-images\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.329614 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e4b586d8-5bf2-482b-8721-e8385575b53a-audit-dir\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.329685 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:31.829657987 +0000 UTC m=+137.810578700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.329825 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61a7f75a-3158-4e5f-a5f9-5c123079888a-trusted-ca\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.329944 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-tls\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.330009 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-trusted-ca-bundle\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.330036 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckfpz\" (UniqueName: \"kubernetes.io/projected/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-kube-api-access-ckfpz\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.330123 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-stats-auth\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.330395 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.331943 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c0c0215-8cde-4704-9ece-1d458f1efd49-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.332117 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-etcd-client\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.332167 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-config\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.332328 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-trusted-ca-bundle\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.332674 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.334068 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-config\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.334157 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61a7f75a-3158-4e5f-a5f9-5c123079888a-trusted-ca\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.334326 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxf9g\" (UniqueName: \"kubernetes.io/projected/37ddf5df-f54d-47f9-8aac-04a2bc93bc25-kube-api-access-kxf9g\") pod \"migrator-59844c95c7-sz52n\" (UID: \"37ddf5df-f54d-47f9-8aac-04a2bc93bc25\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.334565 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.334748 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.334835 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.334902 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-audit-policies\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.335691 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-serving-cert\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.335740 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqz2v\" (UniqueName: \"kubernetes.io/projected/e4b586d8-5bf2-482b-8721-e8385575b53a-kube-api-access-jqz2v\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.335776 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfqq7\" (UniqueName: \"kubernetes.io/projected/4b08791c-eff7-4b2b-8d98-f37e0bdbe887-kube-api-access-nfqq7\") pod \"multus-admission-controller-857f4d67dd-b2d5g\" (UID: \"4b08791c-eff7-4b2b-8d98-f37e0bdbe887\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.336407 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-service-ca\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.336783 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.337486 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.337549 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.337646 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-service-ca\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.337862 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.338013 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-audit-policies\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.338460 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-client-ca\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.339936 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c077bf27-95bf-4abb-b285-8a62e2118c65-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jlxcw\" (UID: \"c077bf27-95bf-4abb-b285-8a62e2118c65\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.340049 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.340175 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4b08791c-eff7-4b2b-8d98-f37e0bdbe887-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-b2d5g\" (UID: \"4b08791c-eff7-4b2b-8d98-f37e0bdbe887\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341119 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-client-ca\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341267 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341330 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341386 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c0c0215-8cde-4704-9ece-1d458f1efd49-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341463 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jvcdk\" (UID: \"2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341532 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bw66\" (UniqueName: \"kubernetes.io/projected/17016565-0937-46c4-bc5c-32b6cd692048-kube-api-access-5bw66\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341571 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqkx4\" (UniqueName: \"kubernetes.io/projected/16dd8be0-522a-4cbc-b837-0562192609b9-kube-api-access-jqkx4\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341664 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da998e2c-9c36-4f73-a839-d81ad59c955d-serving-cert\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341708 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-service-ca\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341750 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61a7f75a-3158-4e5f-a5f9-5c123079888a-metrics-tls\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341820 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341899 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-config\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341937 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxfs6\" (UniqueName: \"kubernetes.io/projected/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-kube-api-access-dxfs6\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.341975 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-oauth-config\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342009 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pb56\" (UniqueName: \"kubernetes.io/projected/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-kube-api-access-9pb56\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342067 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6lvq\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-kube-api-access-t6lvq\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342059 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c0c0215-8cde-4704-9ece-1d458f1efd49-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342099 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rkn8\" (UniqueName: \"kubernetes.io/projected/c077bf27-95bf-4abb-b285-8a62e2118c65-kube-api-access-6rkn8\") pod \"package-server-manager-789f6589d5-jlxcw\" (UID: \"c077bf27-95bf-4abb-b285-8a62e2118c65\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342140 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-config\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342204 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8ltq\" (UniqueName: \"kubernetes.io/projected/2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3-kube-api-access-f8ltq\") pod \"control-plane-machine-set-operator-78cbb6b69f-jvcdk\" (UID: \"2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342240 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-csi-data-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342284 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-config\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342320 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342347 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf9b1cbd-2048-4999-860c-5fab190dd3fb-serving-cert\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342402 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342426 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fpgm\" (UniqueName: \"kubernetes.io/projected/23e81a05-bdfd-4a30-82b0-9eded242e91c-kube-api-access-2fpgm\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342450 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61a7f75a-3158-4e5f-a5f9-5c123079888a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342478 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-trusted-ca\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342499 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-config\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342543 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-bound-sa-token\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342571 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tl2k\" (UniqueName: \"kubernetes.io/projected/bf9b1cbd-2048-4999-860c-5fab190dd3fb-kube-api-access-8tl2k\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342600 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e4b586d8-5bf2-482b-8721-e8385575b53a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.342607 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.343879 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-encryption-config\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.344647 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-trusted-ca\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.345323 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-service-ca\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.346561 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-config\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.346613 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-config\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.346744 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6wtw\" (UniqueName: \"kubernetes.io/projected/61a7f75a-3158-4e5f-a5f9-5c123079888a-kube-api-access-x6wtw\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.346790 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-service-ca-bundle\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.346831 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-metrics-certs\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.346894 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2d188da8-708a-4cdb-a311-39896852177d-metrics-tls\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.347714 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:31.847684892 +0000 UTC m=+137.828605325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.347707 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-config\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.347735 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-certificates\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.348810 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-service-ca-bundle\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.349126 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-certificates\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.349272 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c499a600-3051-43fd-b1e8-24bc9f84f38f-serving-cert\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.350030 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.350968 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf9b1cbd-2048-4999-860c-5fab190dd3fb-config\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.354733 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-default-certificate\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.356410 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-serving-cert\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.356577 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-serving-cert\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.358555 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da998e2c-9c36-4f73-a839-d81ad59c955d-serving-cert\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.359154 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/23e81a05-bdfd-4a30-82b0-9eded242e91c-secret-volume\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.362308 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-metrics-certs\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.362312 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-etcd-client\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.362797 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-stats-auth\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.363777 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-serving-cert\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.364862 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf9b1cbd-2048-4999-860c-5fab190dd3fb-serving-cert\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.365000 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4b08791c-eff7-4b2b-8d98-f37e0bdbe887-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-b2d5g\" (UID: \"4b08791c-eff7-4b2b-8d98-f37e0bdbe887\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.365194 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4b586d8-5bf2-482b-8721-e8385575b53a-etcd-client\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.365418 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-tls\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.366510 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.366818 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jvcdk\" (UID: \"2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.367337 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.367967 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61a7f75a-3158-4e5f-a5f9-5c123079888a-metrics-tls\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.368095 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c499a600-3051-43fd-b1e8-24bc9f84f38f-serving-cert\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.368536 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c077bf27-95bf-4abb-b285-8a62e2118c65-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jlxcw\" (UID: \"c077bf27-95bf-4abb-b285-8a62e2118c65\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.369124 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.374302 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-oauth-config\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.375733 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tkw6\" (UniqueName: \"kubernetes.io/projected/fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26-kube-api-access-8tkw6\") pod \"machine-api-operator-5694c8668f-jldld\" (UID: \"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.406509 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-689m7\" (UniqueName: \"kubernetes.io/projected/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-kube-api-access-689m7\") pod \"marketplace-operator-79b997595-bhp9h\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.426569 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nds6\" (UniqueName: \"kubernetes.io/projected/c499a600-3051-43fd-b1e8-24bc9f84f38f-kube-api-access-8nds6\") pod \"service-ca-operator-777779d784-4pql2\" (UID: \"c499a600-3051-43fd-b1e8-24bc9f84f38f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.445928 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdbx9\" (UniqueName: \"kubernetes.io/projected/da998e2c-9c36-4f73-a839-d81ad59c955d-kube-api-access-mdbx9\") pod \"route-controller-manager-6576b87f9c-d5788\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453102 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453412 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-registration-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453452 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-mountpoint-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453479 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/16dd8be0-522a-4cbc-b837-0562192609b9-node-bootstrap-token\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453509 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d188da8-708a-4cdb-a311-39896852177d-config-volume\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453625 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bw66\" (UniqueName: \"kubernetes.io/projected/17016565-0937-46c4-bc5c-32b6cd692048-kube-api-access-5bw66\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453650 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqkx4\" (UniqueName: \"kubernetes.io/projected/16dd8be0-522a-4cbc-b837-0562192609b9-kube-api-access-jqkx4\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453719 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-csi-data-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453807 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2d188da8-708a-4cdb-a311-39896852177d-metrics-tls\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453855 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvbkn\" (UniqueName: \"kubernetes.io/projected/2d188da8-708a-4cdb-a311-39896852177d-kube-api-access-cvbkn\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453905 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d761d3d-9962-4334-916a-dcaf73bfc51f-cert\") pod \"ingress-canary-ldbtd\" (UID: \"4d761d3d-9962-4334-916a-dcaf73bfc51f\") " pod="openshift-ingress-canary/ingress-canary-ldbtd" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453966 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kczns\" (UniqueName: \"kubernetes.io/projected/4d761d3d-9962-4334-916a-dcaf73bfc51f-kube-api-access-kczns\") pod \"ingress-canary-ldbtd\" (UID: \"4d761d3d-9962-4334-916a-dcaf73bfc51f\") " pod="openshift-ingress-canary/ingress-canary-ldbtd" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.453988 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/16dd8be0-522a-4cbc-b837-0562192609b9-certs\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.454008 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-socket-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.454028 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-plugins-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.454389 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-plugins-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.454493 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:31.954470435 +0000 UTC m=+137.935390868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.454539 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-registration-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.454583 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-mountpoint-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.459220 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-socket-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.459794 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d761d3d-9962-4334-916a-dcaf73bfc51f-cert\") pod \"ingress-canary-ldbtd\" (UID: \"4d761d3d-9962-4334-916a-dcaf73bfc51f\") " pod="openshift-ingress-canary/ingress-canary-ldbtd" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.460067 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/17016565-0937-46c4-bc5c-32b6cd692048-csi-data-dir\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.460115 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d188da8-708a-4cdb-a311-39896852177d-config-volume\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.460635 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/16dd8be0-522a-4cbc-b837-0562192609b9-certs\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.466917 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/16dd8be0-522a-4cbc-b837-0562192609b9-node-bootstrap-token\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.468478 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2d188da8-708a-4cdb-a311-39896852177d-metrics-tls\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.469917 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckfpz\" (UniqueName: \"kubernetes.io/projected/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-kube-api-access-ckfpz\") pod \"console-f9d7485db-sqp4x\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.471084 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.472518 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.473670 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w5nc4"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.486171 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxf9g\" (UniqueName: \"kubernetes.io/projected/37ddf5df-f54d-47f9-8aac-04a2bc93bc25-kube-api-access-kxf9g\") pod \"migrator-59844c95c7-sz52n\" (UID: \"37ddf5df-f54d-47f9-8aac-04a2bc93bc25\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.492690 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.509400 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj9fm\" (UniqueName: \"kubernetes.io/projected/1c0c0215-8cde-4704-9ece-1d458f1efd49-kube-api-access-kj9fm\") pod \"kube-storage-version-migrator-operator-b67b599dd-pf6dq\" (UID: \"1c0c0215-8cde-4704-9ece-1d458f1efd49\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.521581 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqz2v\" (UniqueName: \"kubernetes.io/projected/e4b586d8-5bf2-482b-8721-e8385575b53a-kube-api-access-jqz2v\") pod \"apiserver-7bbb656c7d-px2tg\" (UID: \"e4b586d8-5bf2-482b-8721-e8385575b53a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.527480 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.547743 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfqq7\" (UniqueName: \"kubernetes.io/projected/4b08791c-eff7-4b2b-8d98-f37e0bdbe887-kube-api-access-nfqq7\") pod \"multus-admission-controller-857f4d67dd-b2d5g\" (UID: \"4b08791c-eff7-4b2b-8d98-f37e0bdbe887\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.550588 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.555800 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.556262 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.056240918 +0000 UTC m=+138.037161401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.565383 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0142b7f7-5cfa-4482-8a84-c9f6cf40593c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n47pl\" (UID: \"0142b7f7-5cfa-4482-8a84-c9f6cf40593c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.567950 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.591005 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.598459 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pb56\" (UniqueName: \"kubernetes.io/projected/3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed-kube-api-access-9pb56\") pod \"etcd-operator-b45778765-9clnc\" (UID: \"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.612017 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.621111 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tl2k\" (UniqueName: \"kubernetes.io/projected/bf9b1cbd-2048-4999-860c-5fab190dd3fb-kube-api-access-8tl2k\") pod \"authentication-operator-69f744f599-4mtxh\" (UID: \"bf9b1cbd-2048-4999-860c-5fab190dd3fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.624986 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxfs6\" (UniqueName: \"kubernetes.io/projected/ddc3b385-4bde-4f79-ad4c-60e9d3104ed4-kube-api-access-dxfs6\") pod \"router-default-5444994796-bl4h4\" (UID: \"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4\") " pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.647580 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61a7f75a-3158-4e5f-a5f9-5c123079888a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.657973 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.658568 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.158529255 +0000 UTC m=+138.139449688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.660130 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.660565 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.160547381 +0000 UTC m=+138.141467814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.671147 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6lvq\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-kube-api-access-t6lvq\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.710649 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-m2l5h"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.711087 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.714499 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rkn8\" (UniqueName: \"kubernetes.io/projected/c077bf27-95bf-4abb-b285-8a62e2118c65-kube-api-access-6rkn8\") pod \"package-server-manager-789f6589d5-jlxcw\" (UID: \"c077bf27-95bf-4abb-b285-8a62e2118c65\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.719128 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.727592 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.733748 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.737201 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-bound-sa-token\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.739543 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fpgm\" (UniqueName: \"kubernetes.io/projected/23e81a05-bdfd-4a30-82b0-9eded242e91c-kube-api-access-2fpgm\") pod \"collect-profiles-29412780-726dk\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.741023 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dsmtf"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.741289 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.745947 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6wtw\" (UniqueName: \"kubernetes.io/projected/61a7f75a-3158-4e5f-a5f9-5c123079888a-kube-api-access-x6wtw\") pod \"ingress-operator-5b745b69d9-p8ws5\" (UID: \"61a7f75a-3158-4e5f-a5f9-5c123079888a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.755054 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.761095 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.761700 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.261682516 +0000 UTC m=+138.242602949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.761820 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.775676 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.777766 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8ltq\" (UniqueName: \"kubernetes.io/projected/2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3-kube-api-access-f8ltq\") pod \"control-plane-machine-set-operator-78cbb6b69f-jvcdk\" (UID: \"2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.788251 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.789051 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kv48s"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.805052 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kczns\" (UniqueName: \"kubernetes.io/projected/4d761d3d-9962-4334-916a-dcaf73bfc51f-kube-api-access-kczns\") pod \"ingress-canary-ldbtd\" (UID: \"4d761d3d-9962-4334-916a-dcaf73bfc51f\") " pod="openshift-ingress-canary/ingress-canary-ldbtd" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.805369 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.814284 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" Dec 03 13:11:31 crc kubenswrapper[4690]: W1203 13:11:31.821012 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcea818d_07a3_4c60_8571_c71af537ca5c.slice/crio-585a9039be964a4d44a197d9066eacd301f691211806ea662cb74ed1d8d66013 WatchSource:0}: Error finding container 585a9039be964a4d44a197d9066eacd301f691211806ea662cb74ed1d8d66013: Status 404 returned error can't find the container with id 585a9039be964a4d44a197d9066eacd301f691211806ea662cb74ed1d8d66013 Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.825378 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvbkn\" (UniqueName: \"kubernetes.io/projected/2d188da8-708a-4cdb-a311-39896852177d-kube-api-access-cvbkn\") pod \"dns-default-46m2s\" (UID: \"2d188da8-708a-4cdb-a311-39896852177d\") " pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.844301 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.844502 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.849854 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bw66\" (UniqueName: \"kubernetes.io/projected/17016565-0937-46c4-bc5c-32b6cd692048-kube-api-access-5bw66\") pod \"csi-hostpathplugin-42cnp\" (UID: \"17016565-0937-46c4-bc5c-32b6cd692048\") " pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: W1203 13:11:31.856730 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71232879_c2e3_4d23_b663_5b3cb4bf1d31.slice/crio-5b6390008fe89c836e4b04e84e06ff5c30fc22405a428e8f7d7a3977782a83e0 WatchSource:0}: Error finding container 5b6390008fe89c836e4b04e84e06ff5c30fc22405a428e8f7d7a3977782a83e0: Status 404 returned error can't find the container with id 5b6390008fe89c836e4b04e84e06ff5c30fc22405a428e8f7d7a3977782a83e0 Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.860290 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.862308 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.862655 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.362629735 +0000 UTC m=+138.343550168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.870634 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.880339 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqkx4\" (UniqueName: \"kubernetes.io/projected/16dd8be0-522a-4cbc-b837-0562192609b9-kube-api-access-jqkx4\") pod \"machine-config-server-vhrgj\" (UID: \"16dd8be0-522a-4cbc-b837-0562192609b9\") " pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.885400 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.904967 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-42cnp" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.923080 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ldbtd" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.932391 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vhrgj" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.940904 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.967497 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:31 crc kubenswrapper[4690]: E1203 13:11:31.968722 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.468704138 +0000 UTC m=+138.449624591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.968795 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl"] Dec 03 13:11:31 crc kubenswrapper[4690]: I1203 13:11:31.984472 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq"] Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.069221 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.069625 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.569612277 +0000 UTC m=+138.550532710 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.131880 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xjjml"] Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.171302 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.171763 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.671740729 +0000 UTC m=+138.652661162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.204139 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jldld"] Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.209967 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4pql2"] Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.232551 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn"] Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.233958 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb"] Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.273157 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.273591 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.773571373 +0000 UTC m=+138.754491806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.374042 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.374204 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.874181813 +0000 UTC m=+138.855102256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.374366 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.374721 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.874710658 +0000 UTC m=+138.855631091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.476141 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.476319 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.976291655 +0000 UTC m=+138.957212088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.476366 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.476836 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:32.97682386 +0000 UTC m=+138.957744293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.578294 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.578628 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.078589093 +0000 UTC m=+139.059509576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.578948 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.579408 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.079386935 +0000 UTC m=+139.060307518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.681594 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.681735 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.181706063 +0000 UTC m=+139.162626496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.681940 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.682288 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.182271499 +0000 UTC m=+139.163191932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.783189 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.783423 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.283404673 +0000 UTC m=+139.264325106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.783477 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.784158 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.283853986 +0000 UTC m=+139.264774419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.884275 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.884467 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.384441425 +0000 UTC m=+139.365361858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.884557 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:32 crc kubenswrapper[4690]: E1203 13:11:32.885149 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.385138845 +0000 UTC m=+139.366059338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:32 crc kubenswrapper[4690]: I1203 13:11:32.989956 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:32.999403 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.499248883 +0000 UTC m=+139.480169356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.100007 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.100313 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.600299126 +0000 UTC m=+139.581219559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.201162 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.201481 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.701463101 +0000 UTC m=+139.682383534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.303295 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.303636 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.803621824 +0000 UTC m=+139.784542257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.373970 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" event={"ID":"349d5d37-327a-48e7-bf3d-51e0cfb79c31","Type":"ContainerStarted","Data":"636ade5866a576a58ffc73c525655ce9db333422a54cd6f42185cd076f4fb7a5"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.374881 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" event={"ID":"e35d957c-079a-484a-9508-2a1def9b682d","Type":"ContainerStarted","Data":"c29b878413dd777626fca19d85d0735547bf6604588708d43a8e50176d399812"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.378060 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" event={"ID":"3458d878-4d3a-4989-abe4-22770cec182e","Type":"ContainerStarted","Data":"9036b93413e2cbe7b8c1ce795f2c4918eff32a188febfbb1c8c22a2d4019a1f2"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.379599 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" event={"ID":"2b122759-56c3-4ee4-81dd-cf055ced8121","Type":"ContainerStarted","Data":"d884edd298603be57f281b154aeceb1c490ed6eeda22817b4a7d9191e7c856d1"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.381563 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" event={"ID":"f566aea3-313b-4311-ab29-491756446da4","Type":"ContainerStarted","Data":"f1ebdee8d642348dd2d40a88d8cd280c3fb6a156802370d11daf14b35a2e47ce"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.384721 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.420501 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.421441 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.421974 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:33.921954811 +0000 UTC m=+139.902875244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.433160 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.440477 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sqp4x"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.449035 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dsmtf" event={"ID":"71232879-c2e3-4d23-b663-5b3cb4bf1d31","Type":"ContainerStarted","Data":"5b6390008fe89c836e4b04e84e06ff5c30fc22405a428e8f7d7a3977782a83e0"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.457914 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.461528 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.476436 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" event={"ID":"dcea818d-07a3-4c60-8571-c71af537ca5c","Type":"ContainerStarted","Data":"585a9039be964a4d44a197d9066eacd301f691211806ea662cb74ed1d8d66013"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.476911 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bhp9h"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.484541 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bl4h4" event={"ID":"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4","Type":"ContainerStarted","Data":"136c189732bcdb18796627e2b180c199ba00b440aa0a9a67e1a97476876baf4c"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.485483 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" event={"ID":"1a2df869-a97d-4371-ae6c-fa30342f98df","Type":"ContainerStarted","Data":"d323248dfa7ab5122ef2d4ab32e44d1c60ca136a8f6ca1ad9896610fdcd0bc60"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.488114 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jdlts"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.488926 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" event={"ID":"32821680-989c-49f5-9b31-b5efed94d3c3","Type":"ContainerStarted","Data":"635da785860e2cbc04776faddf3f2be8ac8df973bcc90ed7fafe92be1ab06fcb"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.493356 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" event={"ID":"0922d35f-705f-48af-823a-15155eecabc9","Type":"ContainerStarted","Data":"ae35787dd4dab58f51aeb53af54dd5a7063ffb3f3b8e5de903836ffe0d9461ee"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.496704 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" event={"ID":"75d357bf-d7e7-43cf-904b-0589a663f831","Type":"ContainerStarted","Data":"a43419138661223ff589f2bb6c4dd84ccb2c7738181efd355f7350e8a81a1257"} Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.497525 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl"] Dec 03 13:11:33 crc kubenswrapper[4690]: W1203 13:11:33.499506 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fbba057_1569_44f3_9eb7_1ce074ae625e.slice/crio-803fa4e1143ed2e3839cab4c092140ab9c16ef7a6bdba1bfe32ff7da2e7a2962 WatchSource:0}: Error finding container 803fa4e1143ed2e3839cab4c092140ab9c16ef7a6bdba1bfe32ff7da2e7a2962: Status 404 returned error can't find the container with id 803fa4e1143ed2e3839cab4c092140ab9c16ef7a6bdba1bfe32ff7da2e7a2962 Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.499568 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" event={"ID":"d5c2b268-0de1-4c88-b439-5292c56044dd","Type":"ContainerStarted","Data":"70f4c4da7dce26de4d5962dce8f7781ffeecb4ddbef91635d525650ef02f0e21"} Dec 03 13:11:33 crc kubenswrapper[4690]: W1203 13:11:33.508547 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd786c95_6f85_4a6e_a7c7_341666f2b478.slice/crio-8d7a3af8dfd63a45b4aa7e375f7b7999b8d1f9dc8f31bddee2538e989abb0ee6 WatchSource:0}: Error finding container 8d7a3af8dfd63a45b4aa7e375f7b7999b8d1f9dc8f31bddee2538e989abb0ee6: Status 404 returned error can't find the container with id 8d7a3af8dfd63a45b4aa7e375f7b7999b8d1f9dc8f31bddee2538e989abb0ee6 Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.525188 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.525544 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.025531734 +0000 UTC m=+140.006452167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: W1203 13:11:33.589401 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod907aeacd_adcc_4d58_b59e_e4d1c45f56e6.slice/crio-94053cb8dacd03d444d128351e6a9c983a07eceff1bf590555b07d8b5fb58a3c WatchSource:0}: Error finding container 94053cb8dacd03d444d128351e6a9c983a07eceff1bf590555b07d8b5fb58a3c: Status 404 returned error can't find the container with id 94053cb8dacd03d444d128351e6a9c983a07eceff1bf590555b07d8b5fb58a3c Dec 03 13:11:33 crc kubenswrapper[4690]: W1203 13:11:33.621703 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60890c93_5ee6_4b32_af02_acc98b37dbd3.slice/crio-818a558ec194f2f3a862b9a40a70fa1db78cbb047bb98854192a7fe888bcb22f WatchSource:0}: Error finding container 818a558ec194f2f3a862b9a40a70fa1db78cbb047bb98854192a7fe888bcb22f: Status 404 returned error can't find the container with id 818a558ec194f2f3a862b9a40a70fa1db78cbb047bb98854192a7fe888bcb22f Dec 03 13:11:33 crc kubenswrapper[4690]: W1203 13:11:33.622257 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda998e2c_9c36_4f73_a839_d81ad59c955d.slice/crio-75664de918e11a42228c2f740e69b9a6134cc7db5210d86a467a0d08cb80b67e WatchSource:0}: Error finding container 75664de918e11a42228c2f740e69b9a6134cc7db5210d86a467a0d08cb80b67e: Status 404 returned error can't find the container with id 75664de918e11a42228c2f740e69b9a6134cc7db5210d86a467a0d08cb80b67e Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.630701 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.631133 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.131112304 +0000 UTC m=+140.112032737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: W1203 13:11:33.697613 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4975e4cf_5be0_40e1_a4b0_2db6ec43c207.slice/crio-8993a0d43ac5a838bc415a07ffc63bdfe6b9829c222f3fb569565093798c5e2d WatchSource:0}: Error finding container 8993a0d43ac5a838bc415a07ffc63bdfe6b9829c222f3fb569565093798c5e2d: Status 404 returned error can't find the container with id 8993a0d43ac5a838bc415a07ffc63bdfe6b9829c222f3fb569565093798c5e2d Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.718591 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9clnc"] Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.732404 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.732769 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.232754692 +0000 UTC m=+140.213675125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.838457 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.839133 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5"] Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.841398 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.341361676 +0000 UTC m=+140.322282109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:33 crc kubenswrapper[4690]: W1203 13:11:33.888101 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e5247cf_2bd6_4a35_a7f2_52f4dcfcf5ed.slice/crio-8ba1ff3d8c5e33f8c669843e7afc70496e2c147f3c64fa837429fb6b15fa1fc6 WatchSource:0}: Error finding container 8ba1ff3d8c5e33f8c669843e7afc70496e2c147f3c64fa837429fb6b15fa1fc6: Status 404 returned error can't find the container with id 8ba1ff3d8c5e33f8c669843e7afc70496e2c147f3c64fa837429fb6b15fa1fc6 Dec 03 13:11:33 crc kubenswrapper[4690]: I1203 13:11:33.940890 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:33 crc kubenswrapper[4690]: E1203 13:11:33.941449 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.44142816 +0000 UTC m=+140.422348593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.044689 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.044851 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.544813868 +0000 UTC m=+140.525734301 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.046415 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.046928 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.546910337 +0000 UTC m=+140.527830770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.126309 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw"] Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.147011 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.147535 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.647515187 +0000 UTC m=+140.628435630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.249475 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.249963 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.749947558 +0000 UTC m=+140.730867991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.350513 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.353144 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.853111299 +0000 UTC m=+140.834031752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.392130 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-b2d5g"] Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.422484 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk"] Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.435912 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-46m2s"] Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.452357 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.452807 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:34.952792313 +0000 UTC m=+140.933712746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.491079 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-42cnp"] Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.517699 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ldbtd"] Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.532760 4690 generic.go:334] "Generic (PLEG): container finished" podID="dcea818d-07a3-4c60-8571-c71af537ca5c" containerID="96ee80704159ed75ab26793167a8b8ea8190122e95b6974eed0cc0624431dd6e" exitCode=0 Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.532950 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" event={"ID":"dcea818d-07a3-4c60-8571-c71af537ca5c","Type":"ContainerDied","Data":"96ee80704159ed75ab26793167a8b8ea8190122e95b6974eed0cc0624431dd6e"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.559462 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.561268 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.061245423 +0000 UTC m=+141.042165856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.563364 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" event={"ID":"60890c93-5ee6-4b32-af02-acc98b37dbd3","Type":"ContainerStarted","Data":"818a558ec194f2f3a862b9a40a70fa1db78cbb047bb98854192a7fe888bcb22f"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.563500 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk"] Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.563644 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xjjml" event={"ID":"cd786c95-6f85-4a6e-a7c7-341666f2b478","Type":"ContainerStarted","Data":"8d7a3af8dfd63a45b4aa7e375f7b7999b8d1f9dc8f31bddee2538e989abb0ee6"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.572988 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" event={"ID":"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9","Type":"ContainerStarted","Data":"3f7cf4adeb60487ed4c9a8de6fef36edce0598d7dbf166c07142c11e85be70c6"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.588515 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4mtxh"] Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.590288 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" event={"ID":"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26","Type":"ContainerStarted","Data":"4dec065f0b62b293c3688652b19b7cd678eac19ac36162cdd4e4ec1514bb28ef"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.651653 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" event={"ID":"d5c2b268-0de1-4c88-b439-5292c56044dd","Type":"ContainerStarted","Data":"0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.663248 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" event={"ID":"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed","Type":"ContainerStarted","Data":"8ba1ff3d8c5e33f8c669843e7afc70496e2c147f3c64fa837429fb6b15fa1fc6"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.663316 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.665372 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.667713 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.167688336 +0000 UTC m=+141.148608769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.673366 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bl4h4" event={"ID":"ddc3b385-4bde-4f79-ad4c-60e9d3104ed4","Type":"ContainerStarted","Data":"19768deceb74c1fc8fef9d9ed6de8b21fa76b066ec5f4cf75dd4ff93f5b5db8f"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.674511 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.730321 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.742902 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:34 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:34 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:34 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.742966 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.744279 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" event={"ID":"3458d878-4d3a-4989-abe4-22770cec182e","Type":"ContainerStarted","Data":"a827de83af1b5728f6c519968f8ef2df64e97ff751ddc8277c42ea01dc0bf4fe"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.769777 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.770269 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.270250281 +0000 UTC m=+141.251170714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.789046 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" event={"ID":"32821680-989c-49f5-9b31-b5efed94d3c3","Type":"ContainerStarted","Data":"0210c18445a3eba34c774d485290cfc62cb26b0314121c832461cff96bcca83e"} Dec 03 13:11:34 crc kubenswrapper[4690]: W1203 13:11:34.793317 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17016565_0937_46c4_bc5c_32b6cd692048.slice/crio-f85e2a9ba288679832a4a675d73c04019638facd492e7c1515ec65c8d6a1f848 WatchSource:0}: Error finding container f85e2a9ba288679832a4a675d73c04019638facd492e7c1515ec65c8d6a1f848: Status 404 returned error can't find the container with id f85e2a9ba288679832a4a675d73c04019638facd492e7c1515ec65c8d6a1f848 Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.801331 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" event={"ID":"349d5d37-327a-48e7-bf3d-51e0cfb79c31","Type":"ContainerStarted","Data":"d68066b417c6187fbe50cf64c64d4c11481d9f6169d2cc26d34690c8001b5434"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.801392 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.806280 4690 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-9c5gx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:5443/healthz\": dial tcp 10.217.0.15:5443: connect: connection refused" start-of-body= Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.806330 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" podUID="349d5d37-327a-48e7-bf3d-51e0cfb79c31" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.15:5443/healthz\": dial tcp 10.217.0.15:5443: connect: connection refused" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.846153 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" event={"ID":"0142b7f7-5cfa-4482-8a84-c9f6cf40593c","Type":"ContainerStarted","Data":"8dfbe251271aed859cddbc937735b824f087f95ba74b52408e25614afb7f69f0"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.853501 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" event={"ID":"37ddf5df-f54d-47f9-8aac-04a2bc93bc25","Type":"ContainerStarted","Data":"0b123655f9741344568891ed90e34afa069ce635b68e7a5f874c0979cf37477b"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.858266 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" event={"ID":"e4b586d8-5bf2-482b-8721-e8385575b53a","Type":"ContainerStarted","Data":"f25ba8afb704788fc7c8620cc0a0725f46bfcc3bc5b8b9e16f40595906f541c9"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.873417 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.876128 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.376105918 +0000 UTC m=+141.357026351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.888533 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" event={"ID":"0922d35f-705f-48af-823a-15155eecabc9","Type":"ContainerStarted","Data":"6a9366fbcf5463113ae1e76ff42b4f15f9b8b1820c9533ae9686c3770d526aa7"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.897852 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.897978 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.897996 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" event={"ID":"1c0c0215-8cde-4704-9ece-1d458f1efd49","Type":"ContainerStarted","Data":"991b68aab35883f584a088e14e1e0fbae8696fc75df2480f2fb1b7588ed76336"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.899069 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vhrgj" event={"ID":"16dd8be0-522a-4cbc-b837-0562192609b9","Type":"ContainerStarted","Data":"58d84b02c28717479e2f11a4f27ca97313581c79ce69921f4e6b9341a65f1773"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.912714 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" event={"ID":"1a2df869-a97d-4371-ae6c-fa30342f98df","Type":"ContainerStarted","Data":"ba95bfe53b89294665d675f9fe7fa2e01a01fcff3c90fd64dbe9b18214c35356"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.914284 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" event={"ID":"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5","Type":"ContainerStarted","Data":"34b97920a43b26a8309e3a17cbdb65740d2d51f5843ed1f8740f4f43122ce560"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.915471 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bl4h4" podStartSLOduration=121.915446121 podStartE2EDuration="2m1.915446121s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:34.911054308 +0000 UTC m=+140.891974741" watchObservedRunningTime="2025-12-03 13:11:34.915446121 +0000 UTC m=+140.896366554" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.920857 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" event={"ID":"1778fa3f-41b1-4a29-a253-6b0be39e11c6","Type":"ContainerStarted","Data":"9c4bd227a669b9dd6ced1786780b1a9af060f22f8d7a55bf03b3df37e17b1774"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.926847 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" event={"ID":"c499a600-3051-43fd-b1e8-24bc9f84f38f","Type":"ContainerStarted","Data":"1092ea93d427784355407bae269d2c33305a58f3f942bcb7fa5f3875f01f19e3"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.928689 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" event={"ID":"907aeacd-adcc-4d58-b59e-e4d1c45f56e6","Type":"ContainerStarted","Data":"94053cb8dacd03d444d128351e6a9c983a07eceff1bf590555b07d8b5fb58a3c"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.943235 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" event={"ID":"da998e2c-9c36-4f73-a839-d81ad59c955d","Type":"ContainerStarted","Data":"32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.943402 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" event={"ID":"da998e2c-9c36-4f73-a839-d81ad59c955d","Type":"ContainerStarted","Data":"75664de918e11a42228c2f740e69b9a6134cc7db5210d86a467a0d08cb80b67e"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.945759 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.965520 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" podStartSLOduration=121.965500314 podStartE2EDuration="2m1.965500314s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:34.962678665 +0000 UTC m=+140.943599108" watchObservedRunningTime="2025-12-03 13:11:34.965500314 +0000 UTC m=+140.946420737" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.977893 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" event={"ID":"2b122759-56c3-4ee4-81dd-cf055ced8121","Type":"ContainerStarted","Data":"cd23a0719cc4ee2aae215fccf98fbe02052589762fab8725440998f2a96ca614"} Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.984271 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.985130 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.485103363 +0000 UTC m=+141.466023796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.985298 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.987049 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" event={"ID":"f566aea3-313b-4311-ab29-491756446da4","Type":"ContainerStarted","Data":"f1fbf77bac0b8f1beb3b7bcf83388f77d2be124b2a531988505b831d33bde842"} Dec 03 13:11:34 crc kubenswrapper[4690]: E1203 13:11:34.988174 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.488152399 +0000 UTC m=+141.469073032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:34 crc kubenswrapper[4690]: I1203 13:11:34.991889 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" podStartSLOduration=121.991856322 podStartE2EDuration="2m1.991856322s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:34.989442945 +0000 UTC m=+140.970363388" watchObservedRunningTime="2025-12-03 13:11:34.991856322 +0000 UTC m=+140.972776755" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.013702 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" event={"ID":"75d357bf-d7e7-43cf-904b-0589a663f831","Type":"ContainerStarted","Data":"6245ec35791a4eedc3d7c53174809de139e2657b1762a5f2dcf326be2ecdb075"} Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.025750 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" event={"ID":"4975e4cf-5be0-40e1-a4b0-2db6ec43c207","Type":"ContainerStarted","Data":"8993a0d43ac5a838bc415a07ffc63bdfe6b9829c222f3fb569565093798c5e2d"} Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.070604 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" event={"ID":"61a7f75a-3158-4e5f-a5f9-5c123079888a","Type":"ContainerStarted","Data":"3fa1b6a3b5628fa39f41e37370408ff0783ceb0cbd315a3a41d27947cdc208b9"} Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.093217 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" podStartSLOduration=122.093197443 podStartE2EDuration="2m2.093197443s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:35.093058599 +0000 UTC m=+141.073979042" watchObservedRunningTime="2025-12-03 13:11:35.093197443 +0000 UTC m=+141.074117876" Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.097445 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.597408821 +0000 UTC m=+141.578329254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.097969 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.098580 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.100305 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.600280981 +0000 UTC m=+141.581201574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.109373 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" event={"ID":"281faa96-ab8c-4fa4-886c-b69f1d818c9e","Type":"ContainerStarted","Data":"f123ccf572418861073d2becd555582bf48b0b08a747a880708bdba5b7074b94"} Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.141197 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" event={"ID":"3fbba057-1569-44f3-9eb7-1ce074ae625e","Type":"ContainerStarted","Data":"803fa4e1143ed2e3839cab4c092140ab9c16ef7a6bdba1bfe32ff7da2e7a2962"} Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.149797 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-98mfv" podStartSLOduration=122.149777819 podStartE2EDuration="2m2.149777819s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:35.140634972 +0000 UTC m=+141.121555405" watchObservedRunningTime="2025-12-03 13:11:35.149777819 +0000 UTC m=+141.130698252" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.167115 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" event={"ID":"e35d957c-079a-484a-9508-2a1def9b682d","Type":"ContainerStarted","Data":"5734312e2a21a06b85a6066d6dd99ebef9a7791d9f40f02a04388b93e047b411"} Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.189888 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sqp4x" event={"ID":"fb8fb4ef-5880-4240-9b1a-d111f40d6e44","Type":"ContainerStarted","Data":"c2e5f4aeaade6389b6410accb890510d20d254b34b5623c9785ba36f0d2fda4b"} Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.199377 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4bjl" podStartSLOduration=122.199348668 podStartE2EDuration="2m2.199348668s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:35.189853322 +0000 UTC m=+141.170773755" watchObservedRunningTime="2025-12-03 13:11:35.199348668 +0000 UTC m=+141.180269101" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.202207 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.222160 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.722091556 +0000 UTC m=+141.703011989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.256627 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4b586d8_5bf2_482b_8721_e8385575b53a.slice/crio-31fb20020469239793892704ad08a51b8311446a53b5ecb0d4cda97da5dd7d13.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.323097 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.323511 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.823498288 +0000 UTC m=+141.804418721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.325144 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.337465 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-m2l5h" podStartSLOduration=121.337437639 podStartE2EDuration="2m1.337437639s" podCreationTimestamp="2025-12-03 13:09:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:35.263304301 +0000 UTC m=+141.244224744" watchObservedRunningTime="2025-12-03 13:11:35.337437639 +0000 UTC m=+141.318358072" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.364923 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" podStartSLOduration=121.364890478 podStartE2EDuration="2m1.364890478s" podCreationTimestamp="2025-12-03 13:09:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:35.338427206 +0000 UTC m=+141.319347639" watchObservedRunningTime="2025-12-03 13:11:35.364890478 +0000 UTC m=+141.345810911" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.397041 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-sqp4x" podStartSLOduration=122.397023169 podStartE2EDuration="2m2.397023169s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:35.395862756 +0000 UTC m=+141.376783189" watchObservedRunningTime="2025-12-03 13:11:35.397023169 +0000 UTC m=+141.377943612" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.423783 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.424487 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:35.924466628 +0000 UTC m=+141.905387061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.462559 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hzthl" podStartSLOduration=122.462534675 podStartE2EDuration="2m2.462534675s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:35.455255561 +0000 UTC m=+141.436176014" watchObservedRunningTime="2025-12-03 13:11:35.462534675 +0000 UTC m=+141.443455108" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.526261 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.526704 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.026689113 +0000 UTC m=+142.007609546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.630374 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.631266 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.131224163 +0000 UTC m=+142.112144596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.740389 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.740719 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.240706742 +0000 UTC m=+142.221627175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.765852 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:35 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:35 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:35 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.765931 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.842094 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.843199 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.343176184 +0000 UTC m=+142.324096617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:35 crc kubenswrapper[4690]: I1203 13:11:35.944784 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:35 crc kubenswrapper[4690]: E1203 13:11:35.945443 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.445416479 +0000 UTC m=+142.426336912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.051511 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.052025 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.552003537 +0000 UTC m=+142.532923970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.154880 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.155264 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.655243711 +0000 UTC m=+142.636164144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.231912 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" event={"ID":"2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3","Type":"ContainerStarted","Data":"07c73ce63a529fedd5c358de8b889e3f060182325f6999f7fc5ad51f9d8c7cd5"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.236494 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" event={"ID":"281faa96-ab8c-4fa4-886c-b69f1d818c9e","Type":"ContainerStarted","Data":"3c6f2f3ac6cbfa9d2b0bc665fa1bc53586b7eb7834b924fc324679304e1d6f0b"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.256631 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.256928 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.75690552 +0000 UTC m=+142.737825953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.257098 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.257406 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.757398984 +0000 UTC m=+142.738319417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.279386 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" event={"ID":"1778fa3f-41b1-4a29-a253-6b0be39e11c6","Type":"ContainerStarted","Data":"5e09cb97c1e16e02787ab24f60b38c3bddac759716ef1ecaca3bfe0b84f1033c"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.290133 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" event={"ID":"907aeacd-adcc-4d58-b59e-e4d1c45f56e6","Type":"ContainerStarted","Data":"010f36b191acc75bd494a7a6e07bf800be99e4c88fac2bb16aefc0ccc0a675fc"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.314048 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-brjbq" podStartSLOduration=123.314023911 podStartE2EDuration="2m3.314023911s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.2726125 +0000 UTC m=+142.253532933" watchObservedRunningTime="2025-12-03 13:11:36.314023911 +0000 UTC m=+142.294944344" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.319367 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lsjpb" podStartSLOduration=123.31933931 podStartE2EDuration="2m3.31933931s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.312991952 +0000 UTC m=+142.293912385" watchObservedRunningTime="2025-12-03 13:11:36.31933931 +0000 UTC m=+142.300259743" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.333811 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-46m2s" event={"ID":"2d188da8-708a-4cdb-a311-39896852177d","Type":"ContainerStarted","Data":"2305c282a0fe7c690248490e16fb887780105fd65e175275f11e7aaf1a8d81a0"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.360176 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.361688 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.861667946 +0000 UTC m=+142.842588389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.424505 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" event={"ID":"1c0c0215-8cde-4704-9ece-1d458f1efd49","Type":"ContainerStarted","Data":"8a9dda9562414ae8ea2d5948588fc92b652b2f54236c027854884c8c2f6369e2"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.456572 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-pf6dq" podStartSLOduration=123.456551316 podStartE2EDuration="2m3.456551316s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.455413664 +0000 UTC m=+142.436334097" watchObservedRunningTime="2025-12-03 13:11:36.456551316 +0000 UTC m=+142.437471749" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.462087 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.465249 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:36.965230399 +0000 UTC m=+142.946150832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.472990 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" event={"ID":"2d8a29fb-3ded-4198-8da3-dd09fc8bfad5","Type":"ContainerStarted","Data":"568ac8799b7b3d6b57e1102f33b4e7b915343189acc75f98ff7c4580b0908615"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.493188 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" event={"ID":"3458d878-4d3a-4989-abe4-22770cec182e","Type":"ContainerStarted","Data":"50db9eef2851caa7a8a6585ed3a8a5db0ed1446d86309496d3c8f77b75920aec"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.518945 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" event={"ID":"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26","Type":"ContainerStarted","Data":"3b5ebc4fcecd71ce2101c99c042492ba8419cd121c8ea5ee3bd95d3bd8d6509a"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.525553 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwh9v" podStartSLOduration=123.525531469 podStartE2EDuration="2m3.525531469s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.525349954 +0000 UTC m=+142.506270387" watchObservedRunningTime="2025-12-03 13:11:36.525531469 +0000 UTC m=+142.506451902" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.527589 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h4snn" podStartSLOduration=123.527581957 podStartE2EDuration="2m3.527581957s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.503536613 +0000 UTC m=+142.484457036" watchObservedRunningTime="2025-12-03 13:11:36.527581957 +0000 UTC m=+142.508502390" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.548234 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dsmtf" event={"ID":"71232879-c2e3-4d23-b663-5b3cb4bf1d31","Type":"ContainerStarted","Data":"8dcf8695ab32e8b4c75af4d6b6b74c5dbcfe9e5d7e937a743bebe770bbae8e77"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.549029 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.558530 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" event={"ID":"c499a600-3051-43fd-b1e8-24bc9f84f38f","Type":"ContainerStarted","Data":"66385a921014e000e50e347058173e110c9abac125d91fad9624f1fb97c2f656"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.566013 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.567277 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.067263079 +0000 UTC m=+143.048183512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.594025 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-dsmtf" podStartSLOduration=123.593992498 podStartE2EDuration="2m3.593992498s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.579394319 +0000 UTC m=+142.560314772" watchObservedRunningTime="2025-12-03 13:11:36.593992498 +0000 UTC m=+142.574912951" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.597175 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" event={"ID":"c077bf27-95bf-4abb-b285-8a62e2118c65","Type":"ContainerStarted","Data":"672040deead43298b2e62c2d432024a4b5e3f13507a266fe349cfad3ef76c11b"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.618459 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vhrgj" event={"ID":"16dd8be0-522a-4cbc-b837-0562192609b9","Type":"ContainerStarted","Data":"83c00410d7fb4ecaee204c1afa3e364e97b70537e52bb1c26c50d88913e9c929"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.631292 4690 generic.go:334] "Generic (PLEG): container finished" podID="e4b586d8-5bf2-482b-8721-e8385575b53a" containerID="31fb20020469239793892704ad08a51b8311446a53b5ecb0d4cda97da5dd7d13" exitCode=0 Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.631422 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" event={"ID":"e4b586d8-5bf2-482b-8721-e8385575b53a","Type":"ContainerDied","Data":"31fb20020469239793892704ad08a51b8311446a53b5ecb0d4cda97da5dd7d13"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.667845 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.670812 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.170797541 +0000 UTC m=+143.151717974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.678222 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-vhrgj" podStartSLOduration=8.678200008 podStartE2EDuration="8.678200008s" podCreationTimestamp="2025-12-03 13:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.676770158 +0000 UTC m=+142.657690591" watchObservedRunningTime="2025-12-03 13:11:36.678200008 +0000 UTC m=+142.659120451" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.679321 4690 generic.go:334] "Generic (PLEG): container finished" podID="3fbba057-1569-44f3-9eb7-1ce074ae625e" containerID="97b075dc10dfe10c010190ec97dd9910bb1450e302547bcbecf1ded7c9beab54" exitCode=0 Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.679417 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" event={"ID":"3fbba057-1569-44f3-9eb7-1ce074ae625e","Type":"ContainerDied","Data":"97b075dc10dfe10c010190ec97dd9910bb1450e302547bcbecf1ded7c9beab54"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.679738 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4pql2" podStartSLOduration=122.679731561 podStartE2EDuration="2m2.679731561s" podCreationTimestamp="2025-12-03 13:09:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.627412695 +0000 UTC m=+142.608333128" watchObservedRunningTime="2025-12-03 13:11:36.679731561 +0000 UTC m=+142.660651994" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.697725 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-dsmtf" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.715970 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" event={"ID":"32821680-989c-49f5-9b31-b5efed94d3c3","Type":"ContainerStarted","Data":"5d23e963f7420fb9dfd1e4a818eaaa19a34fb03bf69d5c679a6534a0f9023782"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.730446 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" event={"ID":"3e5247cf-2bd6-4a35-a7f2-52f4dcfcf5ed","Type":"ContainerStarted","Data":"3262852972f1c6ed55cee2d0e1a14a5d511f6a7d2d555252785b22dfa45a630b"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.739096 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:36 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:36 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:36 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.739159 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.773938 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.780337 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" event={"ID":"bf9b1cbd-2048-4999-860c-5fab190dd3fb","Type":"ContainerStarted","Data":"1cf134898c5e7017eb9dca2654e91b80905129f173b54e6cef6e9b6b014c32c2"} Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.776715 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.276676189 +0000 UTC m=+143.257596772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.802111 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" event={"ID":"4975e4cf-5be0-40e1-a4b0-2db6ec43c207","Type":"ContainerStarted","Data":"02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.803348 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.844670 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.871936 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" event={"ID":"23e81a05-bdfd-4a30-82b0-9eded242e91c","Type":"ContainerStarted","Data":"4a0e7c860eab43523cd0d6cf9de8b3710bce3c71765676999ba104c983dfcad5"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.877692 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.885695 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.385672124 +0000 UTC m=+143.366592557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.895809 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" event={"ID":"37ddf5df-f54d-47f9-8aac-04a2bc93bc25","Type":"ContainerStarted","Data":"1f1bd87c8b513445920c90a3ef248cc52348055be8058324b1fed88d08e2a67c"} Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.947290 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" podStartSLOduration=123.94726508 podStartE2EDuration="2m3.94726508s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.90981023 +0000 UTC m=+142.890730663" watchObservedRunningTime="2025-12-03 13:11:36.94726508 +0000 UTC m=+142.928185513" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.952204 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w76dr" podStartSLOduration=123.952171377 podStartE2EDuration="2m3.952171377s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:36.948376871 +0000 UTC m=+142.929297304" watchObservedRunningTime="2025-12-03 13:11:36.952171377 +0000 UTC m=+142.933091810" Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.978667 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:36 crc kubenswrapper[4690]: E1203 13:11:36.980286 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.480262165 +0000 UTC m=+143.461182598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:36 crc kubenswrapper[4690]: I1203 13:11:36.995361 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" event={"ID":"4b08791c-eff7-4b2b-8d98-f37e0bdbe887","Type":"ContainerStarted","Data":"9d87aefa323b32b4fc4360072a1477bd9a961c694c51db0c2e835ff2a8123fe9"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.031101 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sqp4x" event={"ID":"fb8fb4ef-5880-4240-9b1a-d111f40d6e44","Type":"ContainerStarted","Data":"bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.078711 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-9clnc" podStartSLOduration=124.078686564 podStartE2EDuration="2m4.078686564s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.017726255 +0000 UTC m=+142.998646678" watchObservedRunningTime="2025-12-03 13:11:37.078686564 +0000 UTC m=+143.059606997" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.080857 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.081372 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.581358938 +0000 UTC m=+143.562279371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.089327 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" event={"ID":"2b122759-56c3-4ee4-81dd-cf055ced8121","Type":"ContainerStarted","Data":"7b4536817f95321e17eef252a209178a2617d6bff88780f04b6b94060d570b18"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.117841 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" event={"ID":"60890c93-5ee6-4b32-af02-acc98b37dbd3","Type":"ContainerStarted","Data":"1de8aa377c6c4ef50e62d5711f0914a162eeacae5fb66f02717533aa4d0356ff"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.119142 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.120988 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" podStartSLOduration=124.120964648 podStartE2EDuration="2m4.120964648s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.079828736 +0000 UTC m=+143.060749169" watchObservedRunningTime="2025-12-03 13:11:37.120964648 +0000 UTC m=+143.101885081" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.127848 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" podStartSLOduration=124.127821121 podStartE2EDuration="2m4.127821121s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.120221598 +0000 UTC m=+143.101142031" watchObservedRunningTime="2025-12-03 13:11:37.127821121 +0000 UTC m=+143.108741554" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.136014 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" event={"ID":"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9","Type":"ContainerStarted","Data":"07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.137071 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.139347 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xjjml" event={"ID":"cd786c95-6f85-4a6e-a7c7-341666f2b478","Type":"ContainerStarted","Data":"334c25a161398f5fa1a92894096a9c689711ed0b1724c45c4f936c8b71020854"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.139918 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xjjml" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.141890 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xjjml container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.141923 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xjjml" podUID="cd786c95-6f85-4a6e-a7c7-341666f2b478" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.142159 4690 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bhp9h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.142175 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" podUID="4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.162226 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.184301 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.185242 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.68522588 +0000 UTC m=+143.666146313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.190975 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n47pl" event={"ID":"0142b7f7-5cfa-4482-8a84-c9f6cf40593c","Type":"ContainerStarted","Data":"568069ecd40e2150241ed15f389ffd0372f51f42008501141a3ee66f9b56e612"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.227749 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" podStartSLOduration=124.227717391 podStartE2EDuration="2m4.227717391s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.192245236 +0000 UTC m=+143.173165669" watchObservedRunningTime="2025-12-03 13:11:37.227717391 +0000 UTC m=+143.208637824" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.250345 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-42cnp" event={"ID":"17016565-0937-46c4-bc5c-32b6cd692048","Type":"ContainerStarted","Data":"f85e2a9ba288679832a4a675d73c04019638facd492e7c1515ec65c8d6a1f848"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.273650 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-xjjml" podStartSLOduration=124.273624757 podStartE2EDuration="2m4.273624757s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.233309927 +0000 UTC m=+143.214230360" watchObservedRunningTime="2025-12-03 13:11:37.273624757 +0000 UTC m=+143.254545190" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.274003 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-69qxx" podStartSLOduration=124.273998978 podStartE2EDuration="2m4.273998978s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.273387101 +0000 UTC m=+143.254307554" watchObservedRunningTime="2025-12-03 13:11:37.273998978 +0000 UTC m=+143.254919411" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.273738 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ldbtd" event={"ID":"4d761d3d-9962-4334-916a-dcaf73bfc51f","Type":"ContainerStarted","Data":"9851f9018c5e0a07e17164df0c8d7a030c022541215eafb3fce1fce1d04eeafc"} Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.285776 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.288432 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.788400512 +0000 UTC m=+143.769321085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.294133 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c5gx" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.339534 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" podStartSLOduration=124.339502303 podStartE2EDuration="2m4.339502303s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.334843502 +0000 UTC m=+143.315763935" watchObservedRunningTime="2025-12-03 13:11:37.339502303 +0000 UTC m=+143.320422736" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.376977 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gknq8"] Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.378732 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.395846 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.397398 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.423318 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-w5nc4" podStartSLOduration=124.423298282 podStartE2EDuration="2m4.423298282s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.396105609 +0000 UTC m=+143.377026042" watchObservedRunningTime="2025-12-03 13:11:37.423298282 +0000 UTC m=+143.404218715" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.423592 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gknq8"] Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.441748 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:37.941726728 +0000 UTC m=+143.922647161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.460966 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-ldbtd" podStartSLOduration=9.460945807 podStartE2EDuration="9.460945807s" podCreationTimestamp="2025-12-03 13:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:37.442711896 +0000 UTC m=+143.423632349" watchObservedRunningTime="2025-12-03 13:11:37.460945807 +0000 UTC m=+143.441866240" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.527164 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f8wnd"] Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.532041 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-utilities\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.532107 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-catalog-content\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.532148 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.532177 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9x9h\" (UniqueName: \"kubernetes.io/projected/c8c81207-0aa4-4489-a0e5-ed680f9436d2-kube-api-access-j9x9h\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.532590 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.032575814 +0000 UTC m=+144.013496247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.534930 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.538891 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.540723 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f8wnd"] Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.635713 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.636310 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-utilities\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.636359 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-catalog-content\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.636386 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-catalog-content\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.636429 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-utilities\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.636445 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9x9h\" (UniqueName: \"kubernetes.io/projected/c8c81207-0aa4-4489-a0e5-ed680f9436d2-kube-api-access-j9x9h\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.636477 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxr4r\" (UniqueName: \"kubernetes.io/projected/48f93164-1544-42e2-82e0-60a681752349-kube-api-access-vxr4r\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.636650 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.136622881 +0000 UTC m=+144.117543314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.637011 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-utilities\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.637092 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-catalog-content\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.713802 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9x9h\" (UniqueName: \"kubernetes.io/projected/c8c81207-0aa4-4489-a0e5-ed680f9436d2-kube-api-access-j9x9h\") pod \"community-operators-gknq8\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.715160 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hrxpz"] Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.716101 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.737306 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-catalog-content\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.737355 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.737376 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-utilities\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.737403 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxr4r\" (UniqueName: \"kubernetes.io/projected/48f93164-1544-42e2-82e0-60a681752349-kube-api-access-vxr4r\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.738556 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-catalog-content\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.738895 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.238883547 +0000 UTC m=+144.219803980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.739232 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-utilities\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.739353 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:37 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:37 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:37 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.739443 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.774345 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrxpz"] Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.789831 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.849465 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.849822 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-utilities\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.849946 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-catalog-content\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.849995 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwsrl\" (UniqueName: \"kubernetes.io/projected/61bb089c-4d85-45a7-ad6a-b16d95511a29-kube-api-access-zwsrl\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.850160 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.350138595 +0000 UTC m=+144.331059028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.916420 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxr4r\" (UniqueName: \"kubernetes.io/projected/48f93164-1544-42e2-82e0-60a681752349-kube-api-access-vxr4r\") pod \"certified-operators-f8wnd\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.952488 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwsrl\" (UniqueName: \"kubernetes.io/projected/61bb089c-4d85-45a7-ad6a-b16d95511a29-kube-api-access-zwsrl\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.952540 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.952592 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-utilities\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.952668 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-catalog-content\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.953117 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-catalog-content\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: E1203 13:11:37.953692 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.453680157 +0000 UTC m=+144.434600590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.953990 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-utilities\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.970446 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n6pw9"] Dec 03 13:11:37 crc kubenswrapper[4690]: I1203 13:11:37.971449 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.013651 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6pw9"] Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.053403 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.053592 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.553572907 +0000 UTC m=+144.534493340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.053908 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-utilities\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.053975 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-catalog-content\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.054015 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.054042 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjbsh\" (UniqueName: \"kubernetes.io/projected/0907470e-4498-4ff5-b6e1-e4be228e581e-kube-api-access-tjbsh\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.054396 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.55438664 +0000 UTC m=+144.535307073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.094450 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwsrl\" (UniqueName: \"kubernetes.io/projected/61bb089c-4d85-45a7-ad6a-b16d95511a29-kube-api-access-zwsrl\") pod \"community-operators-hrxpz\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.155125 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.155409 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.655384531 +0000 UTC m=+144.636304964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.155843 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-utilities\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.155903 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-catalog-content\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.155927 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.155947 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjbsh\" (UniqueName: \"kubernetes.io/projected/0907470e-4498-4ff5-b6e1-e4be228e581e-kube-api-access-tjbsh\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.156569 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-utilities\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.156783 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-catalog-content\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.157114 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.657106229 +0000 UTC m=+144.638026662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.178404 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.232724 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjbsh\" (UniqueName: \"kubernetes.io/projected/0907470e-4498-4ff5-b6e1-e4be228e581e-kube-api-access-tjbsh\") pod \"certified-operators-n6pw9\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.258382 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.258972 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.758951074 +0000 UTC m=+144.739871507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.305523 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.344890 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" event={"ID":"c077bf27-95bf-4abb-b285-8a62e2118c65","Type":"ContainerStarted","Data":"6b0d8f2bd41fd217ea37dc50bf0c53cbd4d4db3862c128b76aa47d5139eee511"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.344948 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" event={"ID":"c077bf27-95bf-4abb-b285-8a62e2118c65","Type":"ContainerStarted","Data":"54b33f7e03fecf930a40d2a4c4e3363a2fef86eae84e4a4e26b84d945c8dc828"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.346163 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.360737 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.361126 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.861111127 +0000 UTC m=+144.842031560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.381117 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" event={"ID":"1a2df869-a97d-4371-ae6c-fa30342f98df","Type":"ContainerStarted","Data":"a6e2539aa73a189300bc8b37948c4626011c531d74bfb7be492d3e627767386e"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.388036 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.429127 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" podStartSLOduration=125.429093313 podStartE2EDuration="2m5.429093313s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.397837227 +0000 UTC m=+144.378757670" watchObservedRunningTime="2025-12-03 13:11:38.429093313 +0000 UTC m=+144.410013746" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.461807 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.470091 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:38.970063811 +0000 UTC m=+144.950984244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.482956 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-42cnp" event={"ID":"17016565-0937-46c4-bc5c-32b6cd692048","Type":"ContainerStarted","Data":"f4ab8b91a964da2c1ccbe762a47a03f39425ce10ea7060b268f3fdfc87b1de5c"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.526834 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" event={"ID":"e4b586d8-5bf2-482b-8721-e8385575b53a","Type":"ContainerStarted","Data":"ded6d64b753f4c4f722bc64b4971c4a2bbbe2a8929292f39ff783f3de7a28016"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.555461 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7q4tf" podStartSLOduration=125.555437444 podStartE2EDuration="2m5.555437444s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.436179401 +0000 UTC m=+144.417099854" watchObservedRunningTime="2025-12-03 13:11:38.555437444 +0000 UTC m=+144.536357887" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.559773 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mtxh" event={"ID":"bf9b1cbd-2048-4999-860c-5fab190dd3fb","Type":"ContainerStarted","Data":"4dcd5623e3a055fcbeda45095687f373e3e3f0accc96049567dd7415907cb059"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.563158 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.563495 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.063478159 +0000 UTC m=+145.044398592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.564327 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" event={"ID":"fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26","Type":"ContainerStarted","Data":"16cef2cd9dc21ab9528b90d9c08f94ce232e58c46c5c18b682033480ed250f17"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.585322 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" event={"ID":"3fbba057-1569-44f3-9eb7-1ce074ae625e","Type":"ContainerStarted","Data":"12428369df64d5d9309023b4f7af98e6e168714c3025534dc1540bfc7ddcb277"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.585639 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.586414 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" podStartSLOduration=125.586385961 podStartE2EDuration="2m5.586385961s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.557173693 +0000 UTC m=+144.538094126" watchObservedRunningTime="2025-12-03 13:11:38.586385961 +0000 UTC m=+144.567306394" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.598684 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-jldld" podStartSLOduration=125.598652015 podStartE2EDuration="2m5.598652015s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.587919954 +0000 UTC m=+144.568840387" watchObservedRunningTime="2025-12-03 13:11:38.598652015 +0000 UTC m=+144.579572448" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.605115 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" event={"ID":"dcea818d-07a3-4c60-8571-c71af537ca5c","Type":"ContainerStarted","Data":"b8097012de14c0ec9132cc2336b2c149a6b4d51dd975f22a91730b885ad86c97"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.605169 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" event={"ID":"dcea818d-07a3-4c60-8571-c71af537ca5c","Type":"ContainerStarted","Data":"7af81f1cf8088253e8864461e1ffc7e0f9bca56f38094a275c0e90f1d91a346d"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.614463 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gknq8"] Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.636712 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" podStartSLOduration=125.636691251 podStartE2EDuration="2m5.636691251s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.633843011 +0000 UTC m=+144.614763444" watchObservedRunningTime="2025-12-03 13:11:38.636691251 +0000 UTC m=+144.617611684" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.666528 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.666752 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.166715223 +0000 UTC m=+145.147635666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.666808 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.668752 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.16874418 +0000 UTC m=+145.149664613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.681256 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" event={"ID":"1778fa3f-41b1-4a29-a253-6b0be39e11c6","Type":"ContainerStarted","Data":"f0768b13aa2b89d33c338c0e6bccf7d7cc97a54dd7b45d4c0ab3e19d5622ab18"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.699887 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" podStartSLOduration=125.699848872 podStartE2EDuration="2m5.699848872s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.698823013 +0000 UTC m=+144.679743466" watchObservedRunningTime="2025-12-03 13:11:38.699848872 +0000 UTC m=+144.680769295" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.715288 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sz52n" event={"ID":"37ddf5df-f54d-47f9-8aac-04a2bc93bc25","Type":"ContainerStarted","Data":"2b41f61e58ff31d96fc6ccfbe60842a647d016d84bfab61c9dadbebc6d70a1de"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.730705 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ldbtd" event={"ID":"4d761d3d-9962-4334-916a-dcaf73bfc51f","Type":"ContainerStarted","Data":"e8b0061f7217c6d96569a99ae037c2ada0c4b5f0cc326f0a393a3707038023f4"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.762289 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-46m2s" event={"ID":"2d188da8-708a-4cdb-a311-39896852177d","Type":"ContainerStarted","Data":"b556f8b8fdfd5df7ea857f4c4451bc1c1d4f461471b29e7f5337d199af06d16c"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.762461 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.762577 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-46m2s" event={"ID":"2d188da8-708a-4cdb-a311-39896852177d","Type":"ContainerStarted","Data":"1dbcb143fc9efb52817e15508383d807b8703b41b40d0e49e82f569389793470"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.768611 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.769046 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.26901846 +0000 UTC m=+145.249939083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.775369 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:38 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:38 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:38 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.775441 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.782965 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" event={"ID":"61a7f75a-3158-4e5f-a5f9-5c123079888a","Type":"ContainerStarted","Data":"7b1880b6c50478087127de28930d6506a47091665a22f21fee6e80a6f4f8fdf3"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.783025 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" event={"ID":"61a7f75a-3158-4e5f-a5f9-5c123079888a","Type":"ContainerStarted","Data":"8c2d7d88f7f49b6114176db2f4d914672139a2668d4453ef3dad34615fd14a54"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.810611 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" event={"ID":"2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3","Type":"ContainerStarted","Data":"c9d6fa179ac0595239d66ce1824dc1d0ec26288b199faabb05109a87eb1b4e3a"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.825036 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-46m2s" podStartSLOduration=10.824998609 podStartE2EDuration="10.824998609s" podCreationTimestamp="2025-12-03 13:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.824202727 +0000 UTC m=+144.805123160" watchObservedRunningTime="2025-12-03 13:11:38.824998609 +0000 UTC m=+144.805919042" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.825703 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bcqd7" podStartSLOduration=125.825695969 podStartE2EDuration="2m5.825695969s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.747707093 +0000 UTC m=+144.728627526" watchObservedRunningTime="2025-12-03 13:11:38.825695969 +0000 UTC m=+144.806616402" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.845268 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" event={"ID":"4b08791c-eff7-4b2b-8d98-f37e0bdbe887","Type":"ContainerStarted","Data":"a56ddd023fd3cbb50cac2186843e0dea01a27aa45f17cb2419e11a5465254ff8"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.845330 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" event={"ID":"4b08791c-eff7-4b2b-8d98-f37e0bdbe887","Type":"ContainerStarted","Data":"731b4515d49a7518f29d8310be4e6caf8012409b7ee7e37569c1d549c1ea9527"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.873609 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.875696 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" event={"ID":"23e81a05-bdfd-4a30-82b0-9eded242e91c","Type":"ContainerStarted","Data":"c78ade4be78d8a2ea8e784f5b1379b4e7bca6e1f4b13b0601e47348a4e41c8e3"} Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.877445 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xjjml container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.877489 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xjjml" podUID="cd786c95-6f85-4a6e-a7c7-341666f2b478" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.878908 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.37888641 +0000 UTC m=+145.359806843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.903026 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p8ws5" podStartSLOduration=125.903002926 podStartE2EDuration="2m5.903002926s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.872633734 +0000 UTC m=+144.853554167" watchObservedRunningTime="2025-12-03 13:11:38.903002926 +0000 UTC m=+144.883923359" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.931351 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.975591 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.976061 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.476030013 +0000 UTC m=+145.456950446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.976179 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:38 crc kubenswrapper[4690]: E1203 13:11:38.983213 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.483193813 +0000 UTC m=+145.464114246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:38 crc kubenswrapper[4690]: I1203 13:11:38.983780 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-b2d5g" podStartSLOduration=125.983760019 podStartE2EDuration="2m5.983760019s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.982292948 +0000 UTC m=+144.963213381" watchObservedRunningTime="2025-12-03 13:11:38.983760019 +0000 UTC m=+144.964680452" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.020993 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jvcdk" podStartSLOduration=126.020964462 podStartE2EDuration="2m6.020964462s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:38.904264121 +0000 UTC m=+144.885184564" watchObservedRunningTime="2025-12-03 13:11:39.020964462 +0000 UTC m=+145.001884895" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.097343 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:39 crc kubenswrapper[4690]: E1203 13:11:39.103364 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.603333851 +0000 UTC m=+145.584254284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.105006 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:39 crc kubenswrapper[4690]: E1203 13:11:39.105747 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.605732728 +0000 UTC m=+145.586653161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.220398 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:39 crc kubenswrapper[4690]: E1203 13:11:39.221098 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.721080511 +0000 UTC m=+145.702000944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.286141 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f8wnd"] Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.329799 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:39 crc kubenswrapper[4690]: E1203 13:11:39.330261 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.830245441 +0000 UTC m=+145.811165884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lds4l" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.437586 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:39 crc kubenswrapper[4690]: E1203 13:11:39.438024 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:11:39.938004871 +0000 UTC m=+145.918925304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.488344 4690 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.505024 4690 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T13:11:39.488369173Z","Handler":null,"Name":""} Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.539360 4690 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.539411 4690 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.540087 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.551054 4690 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.551097 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.652478 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6pw9"] Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.677254 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrxpz"] Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.755424 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rvd94"] Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.757117 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.758005 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:39 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:39 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:39 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.758044 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.767917 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvd94"] Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.772452 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.842888 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dnj7\" (UniqueName: \"kubernetes.io/projected/c06438b0-cd29-4387-96fd-257f01c8fd11-kube-api-access-6dnj7\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.842975 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-catalog-content\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.843055 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-utilities\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.875601 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lds4l\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.922229 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.926526 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-42cnp" event={"ID":"17016565-0937-46c4-bc5c-32b6cd692048","Type":"ContainerStarted","Data":"024e687e430b8dde4d4d5fd6a1b115fa903492ed7425f07b68b7e0420f5ddf2e"} Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.935617 4690 generic.go:334] "Generic (PLEG): container finished" podID="48f93164-1544-42e2-82e0-60a681752349" containerID="9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879" exitCode=0 Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.935699 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8wnd" event={"ID":"48f93164-1544-42e2-82e0-60a681752349","Type":"ContainerDied","Data":"9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879"} Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.935735 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8wnd" event={"ID":"48f93164-1544-42e2-82e0-60a681752349","Type":"ContainerStarted","Data":"0207cbbe10e776c52ffd3de95108219d95d3f7b3af03019fc7d8d01a063c73e8"} Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.937270 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.947668 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.947856 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-catalog-content\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.947950 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-utilities\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.948024 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dnj7\" (UniqueName: \"kubernetes.io/projected/c06438b0-cd29-4387-96fd-257f01c8fd11-kube-api-access-6dnj7\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.948732 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-catalog-content\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.948967 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-utilities\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.952719 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrxpz" event={"ID":"61bb089c-4d85-45a7-ad6a-b16d95511a29","Type":"ContainerStarted","Data":"9e2f9e5fb7b1e947564d0a375ea876a6d437ec0dc8cb1b3326b36e7fe753ac36"} Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.954437 4690 generic.go:334] "Generic (PLEG): container finished" podID="23e81a05-bdfd-4a30-82b0-9eded242e91c" containerID="c78ade4be78d8a2ea8e784f5b1379b4e7bca6e1f4b13b0601e47348a4e41c8e3" exitCode=0 Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.954498 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" event={"ID":"23e81a05-bdfd-4a30-82b0-9eded242e91c","Type":"ContainerDied","Data":"c78ade4be78d8a2ea8e784f5b1379b4e7bca6e1f4b13b0601e47348a4e41c8e3"} Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.962691 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.963445 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.966716 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6pw9" event={"ID":"0907470e-4498-4ff5-b6e1-e4be228e581e","Type":"ContainerStarted","Data":"740aed24885ddc7375a867daf36857c3fe8f074a0eea4fb2b0af6ca0e07dec14"} Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.993600 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dnj7\" (UniqueName: \"kubernetes.io/projected/c06438b0-cd29-4387-96fd-257f01c8fd11-kube-api-access-6dnj7\") pod \"redhat-marketplace-rvd94\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:39 crc kubenswrapper[4690]: I1203 13:11:39.999569 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.002815 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.002975 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.009338 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerID="bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508" exitCode=0 Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.011216 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gknq8" event={"ID":"c8c81207-0aa4-4489-a0e5-ed680f9436d2","Type":"ContainerDied","Data":"bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508"} Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.011255 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gknq8" event={"ID":"c8c81207-0aa4-4489-a0e5-ed680f9436d2","Type":"ContainerStarted","Data":"ec3c9fa9755df3759defde07001b3e8b9b31368610515149e5893845921de9ac"} Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.012902 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xjjml container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.012965 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xjjml" podUID="cd786c95-6f85-4a6e-a7c7-341666f2b478" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.049997 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.050894 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.050948 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.154112 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-27bwv"] Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.154474 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.154708 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.156062 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.172258 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.183119 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-27bwv"] Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.212760 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.253222 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.256942 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-utilities\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.257216 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-catalog-content\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.257263 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfdtz\" (UniqueName: \"kubernetes.io/projected/b710c0e8-81a2-427b-a964-d80b57047be3-kube-api-access-qfdtz\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.342980 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.343697 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.359151 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-utilities\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.359195 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.359215 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-catalog-content\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.359256 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfdtz\" (UniqueName: \"kubernetes.io/projected/b710c0e8-81a2-427b-a964-d80b57047be3-kube-api-access-qfdtz\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.359276 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.360434 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.360803 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-catalog-content\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.361492 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-utilities\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.364060 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.411657 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfdtz\" (UniqueName: \"kubernetes.io/projected/b710c0e8-81a2-427b-a964-d80b57047be3-kube-api-access-qfdtz\") pod \"redhat-marketplace-27bwv\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.429667 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lds4l"] Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.460767 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.460814 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.464397 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.477452 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.486819 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5j9q8"] Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.488138 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.497951 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.504581 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5j9q8"] Dec 03 13:11:40 crc kubenswrapper[4690]: W1203 13:11:40.511183 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b8dc7b6_5e28_49c8_a0d9_4c17340d579a.slice/crio-88806e92948e40dabb06baa0be3e96958d30ed1fd79476a114e1941c67d289b9 WatchSource:0}: Error finding container 88806e92948e40dabb06baa0be3e96958d30ed1fd79476a114e1941c67d289b9: Status 404 returned error can't find the container with id 88806e92948e40dabb06baa0be3e96958d30ed1fd79476a114e1941c67d289b9 Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.547548 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z8cfl" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.561672 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-catalog-content\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.561758 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-utilities\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.561804 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2w69\" (UniqueName: \"kubernetes.io/projected/9227b45f-65e8-4ed0-877a-fa41f765078f-kube-api-access-h2w69\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.584500 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.635935 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.677223 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2w69\" (UniqueName: \"kubernetes.io/projected/9227b45f-65e8-4ed0-877a-fa41f765078f-kube-api-access-h2w69\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.677313 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-catalog-content\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.677359 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-utilities\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.678504 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-utilities\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.679225 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-catalog-content\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.710450 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2w69\" (UniqueName: \"kubernetes.io/projected/9227b45f-65e8-4ed0-877a-fa41f765078f-kube-api-access-h2w69\") pod \"redhat-operators-5j9q8\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.731426 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.738784 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.740009 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:40 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:40 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:40 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.740041 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.744081 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-drrqc"] Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.745719 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.751052 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-drrqc"] Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.814315 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.879231 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-catalog-content\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.879569 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g77k\" (UniqueName: \"kubernetes.io/projected/6a9984da-231b-45bf-8583-6394924d2e3e-kube-api-access-5g77k\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.879594 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-utilities\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.931747 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvd94"] Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.984288 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-catalog-content\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.984346 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g77k\" (UniqueName: \"kubernetes.io/projected/6a9984da-231b-45bf-8583-6394924d2e3e-kube-api-access-5g77k\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.984373 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-utilities\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.985444 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-utilities\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:40 crc kubenswrapper[4690]: I1203 13:11:40.985960 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-catalog-content\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.019362 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g77k\" (UniqueName: \"kubernetes.io/projected/6a9984da-231b-45bf-8583-6394924d2e3e-kube-api-access-5g77k\") pod \"redhat-operators-drrqc\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.028126 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvd94" event={"ID":"c06438b0-cd29-4387-96fd-257f01c8fd11","Type":"ContainerStarted","Data":"710b57da024e3184f624f9c18088c667c5bb17a794d4dff44edad1b4f5f43eb7"} Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.077349 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-42cnp" event={"ID":"17016565-0937-46c4-bc5c-32b6cd692048","Type":"ContainerStarted","Data":"de6222ba9f9b63bb5bcf59a1d3d4b6c97a64ebf42507a2b62cc3c729a8c080bb"} Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.077815 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-42cnp" event={"ID":"17016565-0937-46c4-bc5c-32b6cd692048","Type":"ContainerStarted","Data":"834fd34d2d9b26e1da77ad5c7bfef623f43f8ffb8952d60a495c310eb1cfaf1e"} Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.090668 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.108454 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.108489 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.112052 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" event={"ID":"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a","Type":"ContainerStarted","Data":"f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9"} Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.112089 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" event={"ID":"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a","Type":"ContainerStarted","Data":"88806e92948e40dabb06baa0be3e96958d30ed1fd79476a114e1941c67d289b9"} Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.113640 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.115065 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-42cnp" podStartSLOduration=13.115054866 podStartE2EDuration="13.115054866s" podCreationTimestamp="2025-12-03 13:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:41.111916768 +0000 UTC m=+147.092837201" watchObservedRunningTime="2025-12-03 13:11:41.115054866 +0000 UTC m=+147.095975299" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.116500 4690 generic.go:334] "Generic (PLEG): container finished" podID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerID="3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733" exitCode=0 Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.116542 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrxpz" event={"ID":"61bb089c-4d85-45a7-ad6a-b16d95511a29","Type":"ContainerDied","Data":"3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733"} Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.133747 4690 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kv48s container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]log ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]etcd ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/max-in-flight-filter ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 13:11:41 crc kubenswrapper[4690]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/openshift.io-startinformers ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 13:11:41 crc kubenswrapper[4690]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 13:11:41 crc kubenswrapper[4690]: livez check failed Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.133799 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" podUID="dcea818d-07a3-4c60-8571-c71af537ca5c" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.147696 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" podStartSLOduration=128.14767848 podStartE2EDuration="2m8.14767848s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:41.145577141 +0000 UTC m=+147.126497574" watchObservedRunningTime="2025-12-03 13:11:41.14767848 +0000 UTC m=+147.128598913" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.170299 4690 generic.go:334] "Generic (PLEG): container finished" podID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerID="b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4" exitCode=0 Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.170436 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6pw9" event={"ID":"0907470e-4498-4ff5-b6e1-e4be228e581e","Type":"ContainerDied","Data":"b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4"} Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.297393 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xjjml container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.297436 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xjjml" podUID="cd786c95-6f85-4a6e-a7c7-341666f2b478" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.297574 4690 patch_prober.go:28] interesting pod/downloads-7954f5f757-xjjml container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.297637 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xjjml" podUID="cd786c95-6f85-4a6e-a7c7-341666f2b478" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.371497 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 13:11:41 crc kubenswrapper[4690]: W1203 13:11:41.429756 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1ddc1f0e_0878_4f0f_b766_750b1df794d5.slice/crio-c8bb45b80282f4075132e9c2943cd83c44081920327ff350c39ebae56e697d2d WatchSource:0}: Error finding container c8bb45b80282f4075132e9c2943cd83c44081920327ff350c39ebae56e697d2d: Status 404 returned error can't find the container with id c8bb45b80282f4075132e9c2943cd83c44081920327ff350c39ebae56e697d2d Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.553663 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.554451 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.564220 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.564236 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.570761 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 13:11:41 crc kubenswrapper[4690]: W1203 13:11:41.619497 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-4382a75423c4fcc196d232b755472b72fbe2869eb0e8971bdba79e19b6ddc0d5 WatchSource:0}: Error finding container 4382a75423c4fcc196d232b755472b72fbe2869eb0e8971bdba79e19b6ddc0d5: Status 404 returned error can't find the container with id 4382a75423c4fcc196d232b755472b72fbe2869eb0e8971bdba79e19b6ddc0d5 Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.656529 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-27bwv"] Dec 03 13:11:41 crc kubenswrapper[4690]: W1203 13:11:41.681239 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-17995397210a951ce9680def7aec98574417f3879bd13d657010aa6c37ac326d WatchSource:0}: Error finding container 17995397210a951ce9680def7aec98574417f3879bd13d657010aa6c37ac326d: Status 404 returned error can't find the container with id 17995397210a951ce9680def7aec98574417f3879bd13d657010aa6c37ac326d Dec 03 13:11:41 crc kubenswrapper[4690]: W1203 13:11:41.697474 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb710c0e8_81a2_427b_a964_d80b57047be3.slice/crio-6cc66ad87fed1f49b80d3281f2bc9cc530e7ca70161d04e6af1ff1f9e834a876 WatchSource:0}: Error finding container 6cc66ad87fed1f49b80d3281f2bc9cc530e7ca70161d04e6af1ff1f9e834a876: Status 404 returned error can't find the container with id 6cc66ad87fed1f49b80d3281f2bc9cc530e7ca70161d04e6af1ff1f9e834a876 Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.697643 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.697699 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.705903 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5j9q8"] Dec 03 13:11:41 crc kubenswrapper[4690]: W1203 13:11:41.706235 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-ea5dbfc73ca96a9013dce11ce0be41635f478bc221849277c04a3f0190ee3c5e WatchSource:0}: Error finding container ea5dbfc73ca96a9013dce11ce0be41635f478bc221849277c04a3f0190ee3c5e: Status 404 returned error can't find the container with id ea5dbfc73ca96a9013dce11ce0be41635f478bc221849277c04a3f0190ee3c5e Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.728374 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.734683 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:41 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:41 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:41 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.734729 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.744521 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.744853 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.754701 4690 patch_prober.go:28] interesting pod/console-f9d7485db-sqp4x container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.754768 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-sqp4x" podUID="fb8fb4ef-5880-4240-9b1a-d111f40d6e44" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.755920 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.756067 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.783045 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.801141 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.801507 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.801989 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.828802 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.960861 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.963724 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:41 crc kubenswrapper[4690]: I1203 13:11:41.969141 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-drrqc"] Dec 03 13:11:42 crc kubenswrapper[4690]: W1203 13:11:42.056332 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a9984da_231b_45bf_8583_6394924d2e3e.slice/crio-108bdf16dcc09a9b654d56b8fca71f24d01a4b13b989a4f146da8c4cd679e87d WatchSource:0}: Error finding container 108bdf16dcc09a9b654d56b8fca71f24d01a4b13b989a4f146da8c4cd679e87d: Status 404 returned error can't find the container with id 108bdf16dcc09a9b654d56b8fca71f24d01a4b13b989a4f146da8c4cd679e87d Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.106858 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/23e81a05-bdfd-4a30-82b0-9eded242e91c-config-volume\") pod \"23e81a05-bdfd-4a30-82b0-9eded242e91c\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.106983 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/23e81a05-bdfd-4a30-82b0-9eded242e91c-secret-volume\") pod \"23e81a05-bdfd-4a30-82b0-9eded242e91c\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.107019 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fpgm\" (UniqueName: \"kubernetes.io/projected/23e81a05-bdfd-4a30-82b0-9eded242e91c-kube-api-access-2fpgm\") pod \"23e81a05-bdfd-4a30-82b0-9eded242e91c\" (UID: \"23e81a05-bdfd-4a30-82b0-9eded242e91c\") " Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.109362 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23e81a05-bdfd-4a30-82b0-9eded242e91c-config-volume" (OuterVolumeSpecName: "config-volume") pod "23e81a05-bdfd-4a30-82b0-9eded242e91c" (UID: "23e81a05-bdfd-4a30-82b0-9eded242e91c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.131645 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e81a05-bdfd-4a30-82b0-9eded242e91c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "23e81a05-bdfd-4a30-82b0-9eded242e91c" (UID: "23e81a05-bdfd-4a30-82b0-9eded242e91c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.149016 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23e81a05-bdfd-4a30-82b0-9eded242e91c-kube-api-access-2fpgm" (OuterVolumeSpecName: "kube-api-access-2fpgm") pod "23e81a05-bdfd-4a30-82b0-9eded242e91c" (UID: "23e81a05-bdfd-4a30-82b0-9eded242e91c"). InnerVolumeSpecName "kube-api-access-2fpgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.209136 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/23e81a05-bdfd-4a30-82b0-9eded242e91c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.209187 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fpgm\" (UniqueName: \"kubernetes.io/projected/23e81a05-bdfd-4a30-82b0-9eded242e91c-kube-api-access-2fpgm\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.209201 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/23e81a05-bdfd-4a30-82b0-9eded242e91c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.229224 4690 generic.go:334] "Generic (PLEG): container finished" podID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerID="7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e" exitCode=0 Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.229775 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvd94" event={"ID":"c06438b0-cd29-4387-96fd-257f01c8fd11","Type":"ContainerDied","Data":"7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.273716 4690 generic.go:334] "Generic (PLEG): container finished" podID="b710c0e8-81a2-427b-a964-d80b57047be3" containerID="62a63967f4fbf8cdf5e267efef15b4da28f6c3917077c6b41ce51b3547d7ba70" exitCode=0 Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.273787 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27bwv" event={"ID":"b710c0e8-81a2-427b-a964-d80b57047be3","Type":"ContainerDied","Data":"62a63967f4fbf8cdf5e267efef15b4da28f6c3917077c6b41ce51b3547d7ba70"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.273814 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27bwv" event={"ID":"b710c0e8-81a2-427b-a964-d80b57047be3","Type":"ContainerStarted","Data":"6cc66ad87fed1f49b80d3281f2bc9cc530e7ca70161d04e6af1ff1f9e834a876"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.280781 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drrqc" event={"ID":"6a9984da-231b-45bf-8583-6394924d2e3e","Type":"ContainerStarted","Data":"108bdf16dcc09a9b654d56b8fca71f24d01a4b13b989a4f146da8c4cd679e87d"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.285994 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ea5dbfc73ca96a9013dce11ce0be41635f478bc221849277c04a3f0190ee3c5e"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.290994 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j9q8" event={"ID":"9227b45f-65e8-4ed0-877a-fa41f765078f","Type":"ContainerStarted","Data":"68ec7708b1166b5a321fcf7b9c1d71c4f4157d3e32d5b10681108b81d8765439"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.299729 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"17995397210a951ce9680def7aec98574417f3879bd13d657010aa6c37ac326d"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.337182 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.347775 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8286fb3a94806ae3ea92f1bb6c5e03c336fcefda71b21af597d2157950d24f08"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.347825 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4382a75423c4fcc196d232b755472b72fbe2869eb0e8971bdba79e19b6ddc0d5"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.347837 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk" event={"ID":"23e81a05-bdfd-4a30-82b0-9eded242e91c","Type":"ContainerDied","Data":"4a0e7c860eab43523cd0d6cf9de8b3710bce3c71765676999ba104c983dfcad5"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.347854 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a0e7c860eab43523cd0d6cf9de8b3710bce3c71765676999ba104c983dfcad5" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.348467 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.353310 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.366798 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1ddc1f0e-0878-4f0f-b766-750b1df794d5","Type":"ContainerStarted","Data":"c8bb45b80282f4075132e9c2943cd83c44081920327ff350c39ebae56e697d2d"} Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.377063 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-px2tg" Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.421831 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.421808502 podStartE2EDuration="3.421808502s" podCreationTimestamp="2025-12-03 13:11:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:42.389790575 +0000 UTC m=+148.370711008" watchObservedRunningTime="2025-12-03 13:11:42.421808502 +0000 UTC m=+148.402728935" Dec 03 13:11:42 crc kubenswrapper[4690]: W1203 13:11:42.450465 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod8b1ab89f_1ed0_4782_a421_4e3ef98d31b8.slice/crio-4bfa3b590b30de55b90a5c6ea2e06df3751c93abdb7395fac7a1c3ef3348ae58 WatchSource:0}: Error finding container 4bfa3b590b30de55b90a5c6ea2e06df3751c93abdb7395fac7a1c3ef3348ae58: Status 404 returned error can't find the container with id 4bfa3b590b30de55b90a5c6ea2e06df3751c93abdb7395fac7a1c3ef3348ae58 Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.738139 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:42 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:42 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:42 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:42 crc kubenswrapper[4690]: I1203 13:11:42.738194 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.395001 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fde625a91283b841f5e1ea03131ed4fd8e6204d39b3a50ca0d642c91ceecba78"} Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.401281 4690 generic.go:334] "Generic (PLEG): container finished" podID="1ddc1f0e-0878-4f0f-b766-750b1df794d5" containerID="ea1536aee038a990a2178fdf289eef09092d37c31dff973c96427959a436159a" exitCode=0 Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.401338 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1ddc1f0e-0878-4f0f-b766-750b1df794d5","Type":"ContainerDied","Data":"ea1536aee038a990a2178fdf289eef09092d37c31dff973c96427959a436159a"} Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.406749 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"497d1d4e108ebaa7cb2253e57de6204f1e21443cc01d7eb06ad233776deb8edd"} Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.415142 4690 generic.go:334] "Generic (PLEG): container finished" podID="6a9984da-231b-45bf-8583-6394924d2e3e" containerID="7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26" exitCode=0 Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.415227 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drrqc" event={"ID":"6a9984da-231b-45bf-8583-6394924d2e3e","Type":"ContainerDied","Data":"7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26"} Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.444456 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8","Type":"ContainerStarted","Data":"34586a6c338f675fbe5edc6818342b7a15598d4ea8ef9da78ff1bff04ce7b3c9"} Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.444510 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8","Type":"ContainerStarted","Data":"4bfa3b590b30de55b90a5c6ea2e06df3751c93abdb7395fac7a1c3ef3348ae58"} Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.462107 4690 generic.go:334] "Generic (PLEG): container finished" podID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerID="ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636" exitCode=0 Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.463218 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j9q8" event={"ID":"9227b45f-65e8-4ed0-877a-fa41f765078f","Type":"ContainerDied","Data":"ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636"} Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.480246 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.480223689 podStartE2EDuration="2.480223689s" podCreationTimestamp="2025-12-03 13:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:11:43.477019819 +0000 UTC m=+149.457940252" watchObservedRunningTime="2025-12-03 13:11:43.480223689 +0000 UTC m=+149.461144112" Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.737706 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:43 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:43 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:43 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:43 crc kubenswrapper[4690]: I1203 13:11:43.737780 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:44 crc kubenswrapper[4690]: I1203 13:11:44.476209 4690 generic.go:334] "Generic (PLEG): container finished" podID="8b1ab89f-1ed0-4782-a421-4e3ef98d31b8" containerID="34586a6c338f675fbe5edc6818342b7a15598d4ea8ef9da78ff1bff04ce7b3c9" exitCode=0 Dec 03 13:11:44 crc kubenswrapper[4690]: I1203 13:11:44.477134 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8","Type":"ContainerDied","Data":"34586a6c338f675fbe5edc6818342b7a15598d4ea8ef9da78ff1bff04ce7b3c9"} Dec 03 13:11:44 crc kubenswrapper[4690]: I1203 13:11:44.736305 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:44 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:44 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:44 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:44 crc kubenswrapper[4690]: I1203 13:11:44.737452 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:44 crc kubenswrapper[4690]: I1203 13:11:44.872901 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:44 crc kubenswrapper[4690]: I1203 13:11:44.999614 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kubelet-dir\") pod \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\" (UID: \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\") " Dec 03 13:11:44 crc kubenswrapper[4690]: I1203 13:11:44.999771 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1ddc1f0e-0878-4f0f-b766-750b1df794d5" (UID: "1ddc1f0e-0878-4f0f-b766-750b1df794d5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:11:44 crc kubenswrapper[4690]: I1203 13:11:44.999805 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kube-api-access\") pod \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\" (UID: \"1ddc1f0e-0878-4f0f-b766-750b1df794d5\") " Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.001346 4690 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.006576 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1ddc1f0e-0878-4f0f-b766-750b1df794d5" (UID: "1ddc1f0e-0878-4f0f-b766-750b1df794d5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.112610 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ddc1f0e-0878-4f0f-b766-750b1df794d5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.294509 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.522959 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.523902 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1ddc1f0e-0878-4f0f-b766-750b1df794d5","Type":"ContainerDied","Data":"c8bb45b80282f4075132e9c2943cd83c44081920327ff350c39ebae56e697d2d"} Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.523977 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8bb45b80282f4075132e9c2943cd83c44081920327ff350c39ebae56e697d2d" Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.733353 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:45 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:45 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:45 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:45 crc kubenswrapper[4690]: I1203 13:11:45.734020 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.109925 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.114328 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kv48s" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.733162 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.733603 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:46 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:46 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:46 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.733650 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.825859 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.825943 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.871118 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8b1ab89f-1ed0-4782-a421-4e3ef98d31b8" (UID: "8b1ab89f-1ed0-4782-a421-4e3ef98d31b8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.871486 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kubelet-dir\") pod \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\" (UID: \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\") " Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.871601 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kube-api-access\") pod \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\" (UID: \"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8\") " Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.872468 4690 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.881988 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8b1ab89f-1ed0-4782-a421-4e3ef98d31b8" (UID: "8b1ab89f-1ed0-4782-a421-4e3ef98d31b8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.945821 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-46m2s" Dec 03 13:11:46 crc kubenswrapper[4690]: I1203 13:11:46.973381 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b1ab89f-1ed0-4782-a421-4e3ef98d31b8-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:47 crc kubenswrapper[4690]: I1203 13:11:47.557510 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8b1ab89f-1ed0-4782-a421-4e3ef98d31b8","Type":"ContainerDied","Data":"4bfa3b590b30de55b90a5c6ea2e06df3751c93abdb7395fac7a1c3ef3348ae58"} Dec 03 13:11:47 crc kubenswrapper[4690]: I1203 13:11:47.557565 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bfa3b590b30de55b90a5c6ea2e06df3751c93abdb7395fac7a1c3ef3348ae58" Dec 03 13:11:47 crc kubenswrapper[4690]: I1203 13:11:47.557677 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:11:47 crc kubenswrapper[4690]: I1203 13:11:47.731098 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:47 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:47 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:47 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:47 crc kubenswrapper[4690]: I1203 13:11:47.731621 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:48 crc kubenswrapper[4690]: I1203 13:11:48.735378 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:48 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:48 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:48 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:48 crc kubenswrapper[4690]: I1203 13:11:48.735450 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:49 crc kubenswrapper[4690]: I1203 13:11:49.731465 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:49 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:49 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:49 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:49 crc kubenswrapper[4690]: I1203 13:11:49.731532 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:50 crc kubenswrapper[4690]: I1203 13:11:50.731138 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:50 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:50 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:50 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:50 crc kubenswrapper[4690]: I1203 13:11:50.731191 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:51 crc kubenswrapper[4690]: I1203 13:11:51.311507 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-xjjml" Dec 03 13:11:51 crc kubenswrapper[4690]: I1203 13:11:51.732787 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:51 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:51 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:51 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:51 crc kubenswrapper[4690]: I1203 13:11:51.732900 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:51 crc kubenswrapper[4690]: I1203 13:11:51.735555 4690 patch_prober.go:28] interesting pod/console-f9d7485db-sqp4x container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 03 13:11:51 crc kubenswrapper[4690]: I1203 13:11:51.735643 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-sqp4x" podUID="fb8fb4ef-5880-4240-9b1a-d111f40d6e44" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 03 13:11:52 crc kubenswrapper[4690]: I1203 13:11:52.731723 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:52 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:52 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:52 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:52 crc kubenswrapper[4690]: I1203 13:11:52.731798 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:53 crc kubenswrapper[4690]: I1203 13:11:53.730310 4690 patch_prober.go:28] interesting pod/router-default-5444994796-bl4h4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:11:53 crc kubenswrapper[4690]: [-]has-synced failed: reason withheld Dec 03 13:11:53 crc kubenswrapper[4690]: [+]process-running ok Dec 03 13:11:53 crc kubenswrapper[4690]: healthz check failed Dec 03 13:11:53 crc kubenswrapper[4690]: I1203 13:11:53.730382 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bl4h4" podUID="ddc3b385-4bde-4f79-ad4c-60e9d3104ed4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:11:54 crc kubenswrapper[4690]: I1203 13:11:54.731715 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:54 crc kubenswrapper[4690]: I1203 13:11:54.734416 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bl4h4" Dec 03 13:11:56 crc kubenswrapper[4690]: I1203 13:11:56.053121 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:56 crc kubenswrapper[4690]: I1203 13:11:56.068238 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abab6c63-91d1-4f36-9be9-ae3b1a43742e-metrics-certs\") pod \"network-metrics-daemon-mbbpr\" (UID: \"abab6c63-91d1-4f36-9be9-ae3b1a43742e\") " pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:56 crc kubenswrapper[4690]: I1203 13:11:56.340454 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mbbpr" Dec 03 13:11:59 crc kubenswrapper[4690]: I1203 13:11:59.928500 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:12:01 crc kubenswrapper[4690]: I1203 13:12:01.739611 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:12:01 crc kubenswrapper[4690]: I1203 13:12:01.744253 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:12:11 crc kubenswrapper[4690]: I1203 13:12:11.811675 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlxcw" Dec 03 13:12:16 crc kubenswrapper[4690]: I1203 13:12:16.824989 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:12:16 crc kubenswrapper[4690]: I1203 13:12:16.825503 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:12:17 crc kubenswrapper[4690]: E1203 13:12:17.054130 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 13:12:17 crc kubenswrapper[4690]: E1203 13:12:17.054666 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zwsrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hrxpz_openshift-marketplace(61bb089c-4d85-45a7-ad6a-b16d95511a29): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:12:17 crc kubenswrapper[4690]: E1203 13:12:17.056136 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hrxpz" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" Dec 03 13:12:17 crc kubenswrapper[4690]: E1203 13:12:17.552964 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 13:12:17 crc kubenswrapper[4690]: E1203 13:12:17.553158 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9x9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gknq8_openshift-marketplace(c8c81207-0aa4-4489-a0e5-ed680f9436d2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:12:17 crc kubenswrapper[4690]: E1203 13:12:17.554382 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gknq8" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" Dec 03 13:12:18 crc kubenswrapper[4690]: E1203 13:12:18.295991 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-gknq8" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" Dec 03 13:12:18 crc kubenswrapper[4690]: E1203 13:12:18.296075 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hrxpz" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" Dec 03 13:12:18 crc kubenswrapper[4690]: E1203 13:12:18.387099 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 13:12:18 crc kubenswrapper[4690]: E1203 13:12:18.387801 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qfdtz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-27bwv_openshift-marketplace(b710c0e8-81a2-427b-a964-d80b57047be3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:12:18 crc kubenswrapper[4690]: E1203 13:12:18.390902 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-27bwv" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" Dec 03 13:12:18 crc kubenswrapper[4690]: E1203 13:12:18.405129 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 13:12:18 crc kubenswrapper[4690]: E1203 13:12:18.405297 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6dnj7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rvd94_openshift-marketplace(c06438b0-cd29-4387-96fd-257f01c8fd11): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:12:18 crc kubenswrapper[4690]: E1203 13:12:18.406708 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rvd94" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" Dec 03 13:12:19 crc kubenswrapper[4690]: E1203 13:12:19.932453 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-27bwv" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" Dec 03 13:12:19 crc kubenswrapper[4690]: E1203 13:12:19.932479 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rvd94" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" Dec 03 13:12:20 crc kubenswrapper[4690]: E1203 13:12:20.013584 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 13:12:20 crc kubenswrapper[4690]: E1203 13:12:20.013778 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 13:12:20 crc kubenswrapper[4690]: E1203 13:12:20.014135 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tjbsh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-n6pw9_openshift-marketplace(0907470e-4498-4ff5-b6e1-e4be228e581e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:12:20 crc kubenswrapper[4690]: E1203 13:12:20.014275 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vxr4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-f8wnd_openshift-marketplace(48f93164-1544-42e2-82e0-60a681752349): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:12:20 crc kubenswrapper[4690]: E1203 13:12:20.015651 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-n6pw9" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" Dec 03 13:12:20 crc kubenswrapper[4690]: E1203 13:12:20.015750 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-f8wnd" podUID="48f93164-1544-42e2-82e0-60a681752349" Dec 03 13:12:20 crc kubenswrapper[4690]: I1203 13:12:20.744107 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.296561 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 13:12:22 crc kubenswrapper[4690]: E1203 13:12:22.297649 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e81a05-bdfd-4a30-82b0-9eded242e91c" containerName="collect-profiles" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.297666 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e81a05-bdfd-4a30-82b0-9eded242e91c" containerName="collect-profiles" Dec 03 13:12:22 crc kubenswrapper[4690]: E1203 13:12:22.297676 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddc1f0e-0878-4f0f-b766-750b1df794d5" containerName="pruner" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.297685 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddc1f0e-0878-4f0f-b766-750b1df794d5" containerName="pruner" Dec 03 13:12:22 crc kubenswrapper[4690]: E1203 13:12:22.297693 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1ab89f-1ed0-4782-a421-4e3ef98d31b8" containerName="pruner" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.297702 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1ab89f-1ed0-4782-a421-4e3ef98d31b8" containerName="pruner" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.297839 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ddc1f0e-0878-4f0f-b766-750b1df794d5" containerName="pruner" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.297859 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="23e81a05-bdfd-4a30-82b0-9eded242e91c" containerName="collect-profiles" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.297875 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1ab89f-1ed0-4782-a421-4e3ef98d31b8" containerName="pruner" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.298703 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.310325 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.310724 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.326249 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.409426 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb636775-ffe4-4dd0-9874-f878fcb9f737-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb636775-ffe4-4dd0-9874-f878fcb9f737\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.409499 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb636775-ffe4-4dd0-9874-f878fcb9f737-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb636775-ffe4-4dd0-9874-f878fcb9f737\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.511733 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb636775-ffe4-4dd0-9874-f878fcb9f737-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb636775-ffe4-4dd0-9874-f878fcb9f737\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.511856 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb636775-ffe4-4dd0-9874-f878fcb9f737-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb636775-ffe4-4dd0-9874-f878fcb9f737\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.512258 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb636775-ffe4-4dd0-9874-f878fcb9f737-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb636775-ffe4-4dd0-9874-f878fcb9f737\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.543821 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb636775-ffe4-4dd0-9874-f878fcb9f737-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb636775-ffe4-4dd0-9874-f878fcb9f737\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:22 crc kubenswrapper[4690]: I1203 13:12:22.639846 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.021566 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-f8wnd" podUID="48f93164-1544-42e2-82e0-60a681752349" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.022635 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-n6pw9" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.104040 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.104259 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h2w69,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5j9q8_openshift-marketplace(9227b45f-65e8-4ed0-877a-fa41f765078f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.105647 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5j9q8" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.127261 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.127481 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5g77k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-drrqc_openshift-marketplace(6a9984da-231b-45bf-8583-6394924d2e3e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.129859 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-drrqc" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" Dec 03 13:12:23 crc kubenswrapper[4690]: I1203 13:12:23.447173 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mbbpr"] Dec 03 13:12:23 crc kubenswrapper[4690]: W1203 13:12:23.453923 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabab6c63_91d1_4f36_9be9_ae3b1a43742e.slice/crio-fce9192c6248f6ee5d0a474ce5e93a4ff0c7d54a97b92e65720676eb728983a9 WatchSource:0}: Error finding container fce9192c6248f6ee5d0a474ce5e93a4ff0c7d54a97b92e65720676eb728983a9: Status 404 returned error can't find the container with id fce9192c6248f6ee5d0a474ce5e93a4ff0c7d54a97b92e65720676eb728983a9 Dec 03 13:12:23 crc kubenswrapper[4690]: I1203 13:12:23.500213 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 13:12:23 crc kubenswrapper[4690]: W1203 13:12:23.507357 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podfb636775_ffe4_4dd0_9874_f878fcb9f737.slice/crio-ec7fea3df7e79f32f9e9657e18112a9b8b0b9c53f4bdab5258af46509b75f5bc WatchSource:0}: Error finding container ec7fea3df7e79f32f9e9657e18112a9b8b0b9c53f4bdab5258af46509b75f5bc: Status 404 returned error can't find the container with id ec7fea3df7e79f32f9e9657e18112a9b8b0b9c53f4bdab5258af46509b75f5bc Dec 03 13:12:23 crc kubenswrapper[4690]: I1203 13:12:23.806737 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" event={"ID":"abab6c63-91d1-4f36-9be9-ae3b1a43742e","Type":"ContainerStarted","Data":"fce9192c6248f6ee5d0a474ce5e93a4ff0c7d54a97b92e65720676eb728983a9"} Dec 03 13:12:23 crc kubenswrapper[4690]: I1203 13:12:23.809044 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb636775-ffe4-4dd0-9874-f878fcb9f737","Type":"ContainerStarted","Data":"ec7fea3df7e79f32f9e9657e18112a9b8b0b9c53f4bdab5258af46509b75f5bc"} Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.809759 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5j9q8" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" Dec 03 13:12:23 crc kubenswrapper[4690]: E1203 13:12:23.810195 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-drrqc" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" Dec 03 13:12:24 crc kubenswrapper[4690]: I1203 13:12:24.817463 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" event={"ID":"abab6c63-91d1-4f36-9be9-ae3b1a43742e","Type":"ContainerStarted","Data":"df45620debf326096ad1d8ff3a612737159f66e2e6284ac61a19e336f8e359a5"} Dec 03 13:12:24 crc kubenswrapper[4690]: I1203 13:12:24.817903 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mbbpr" event={"ID":"abab6c63-91d1-4f36-9be9-ae3b1a43742e","Type":"ContainerStarted","Data":"5c09bf3fe03806a54b181312fb5107585028ac4c6649ba5967599a82c76abb5f"} Dec 03 13:12:24 crc kubenswrapper[4690]: I1203 13:12:24.820344 4690 generic.go:334] "Generic (PLEG): container finished" podID="fb636775-ffe4-4dd0-9874-f878fcb9f737" containerID="2463fba14e5fd196d2028f684777e9507d5b47c723a489ddd7098b709664954b" exitCode=0 Dec 03 13:12:24 crc kubenswrapper[4690]: I1203 13:12:24.820397 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb636775-ffe4-4dd0-9874-f878fcb9f737","Type":"ContainerDied","Data":"2463fba14e5fd196d2028f684777e9507d5b47c723a489ddd7098b709664954b"} Dec 03 13:12:24 crc kubenswrapper[4690]: I1203 13:12:24.837438 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-mbbpr" podStartSLOduration=171.837410266 podStartE2EDuration="2m51.837410266s" podCreationTimestamp="2025-12-03 13:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:12:24.833532328 +0000 UTC m=+190.814452761" watchObservedRunningTime="2025-12-03 13:12:24.837410266 +0000 UTC m=+190.818330719" Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.073841 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.165000 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb636775-ffe4-4dd0-9874-f878fcb9f737-kubelet-dir\") pod \"fb636775-ffe4-4dd0-9874-f878fcb9f737\" (UID: \"fb636775-ffe4-4dd0-9874-f878fcb9f737\") " Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.165072 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb636775-ffe4-4dd0-9874-f878fcb9f737-kube-api-access\") pod \"fb636775-ffe4-4dd0-9874-f878fcb9f737\" (UID: \"fb636775-ffe4-4dd0-9874-f878fcb9f737\") " Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.165151 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb636775-ffe4-4dd0-9874-f878fcb9f737-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fb636775-ffe4-4dd0-9874-f878fcb9f737" (UID: "fb636775-ffe4-4dd0-9874-f878fcb9f737"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.165305 4690 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb636775-ffe4-4dd0-9874-f878fcb9f737-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.170570 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb636775-ffe4-4dd0-9874-f878fcb9f737-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fb636775-ffe4-4dd0-9874-f878fcb9f737" (UID: "fb636775-ffe4-4dd0-9874-f878fcb9f737"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.266698 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb636775-ffe4-4dd0-9874-f878fcb9f737-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.835507 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb636775-ffe4-4dd0-9874-f878fcb9f737","Type":"ContainerDied","Data":"ec7fea3df7e79f32f9e9657e18112a9b8b0b9c53f4bdab5258af46509b75f5bc"} Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.835551 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec7fea3df7e79f32f9e9657e18112a9b8b0b9c53f4bdab5258af46509b75f5bc" Dec 03 13:12:26 crc kubenswrapper[4690]: I1203 13:12:26.835611 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.696664 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 13:12:28 crc kubenswrapper[4690]: E1203 13:12:28.698225 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb636775-ffe4-4dd0-9874-f878fcb9f737" containerName="pruner" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.698354 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb636775-ffe4-4dd0-9874-f878fcb9f737" containerName="pruner" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.698569 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb636775-ffe4-4dd0-9874-f878fcb9f737" containerName="pruner" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.699096 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.701148 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.701276 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.707423 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.799051 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kubelet-dir\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.799404 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kube-api-access\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.799573 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-var-lock\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.901261 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kube-api-access\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.901368 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-var-lock\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.901463 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kubelet-dir\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.901537 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-var-lock\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.901583 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kubelet-dir\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:28 crc kubenswrapper[4690]: I1203 13:12:28.919540 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kube-api-access\") pod \"installer-9-crc\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:29 crc kubenswrapper[4690]: I1203 13:12:29.017592 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:12:29 crc kubenswrapper[4690]: I1203 13:12:29.404605 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 13:12:29 crc kubenswrapper[4690]: I1203 13:12:29.859604 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"455990f2-8fc3-4aa8-9a20-1c1f4d747d51","Type":"ContainerStarted","Data":"852a183a677ab4500648ed661080ab91234e747362104ae04d4cb1c22865308d"} Dec 03 13:12:29 crc kubenswrapper[4690]: I1203 13:12:29.859653 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"455990f2-8fc3-4aa8-9a20-1c1f4d747d51","Type":"ContainerStarted","Data":"8126d0a1726b25c140029123d5359015ce6deedb14ad3079cc5111c8ce7cf64d"} Dec 03 13:12:29 crc kubenswrapper[4690]: I1203 13:12:29.880544 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.880525697 podStartE2EDuration="1.880525697s" podCreationTimestamp="2025-12-03 13:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:12:29.879597431 +0000 UTC m=+195.860517884" watchObservedRunningTime="2025-12-03 13:12:29.880525697 +0000 UTC m=+195.861446140" Dec 03 13:12:29 crc kubenswrapper[4690]: I1203 13:12:29.917846 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54f5g"] Dec 03 13:12:34 crc kubenswrapper[4690]: I1203 13:12:34.897199 4690 generic.go:334] "Generic (PLEG): container finished" podID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerID="412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8" exitCode=0 Dec 03 13:12:34 crc kubenswrapper[4690]: I1203 13:12:34.897288 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrxpz" event={"ID":"61bb089c-4d85-45a7-ad6a-b16d95511a29","Type":"ContainerDied","Data":"412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8"} Dec 03 13:12:34 crc kubenswrapper[4690]: I1203 13:12:34.901720 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerID="6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e" exitCode=0 Dec 03 13:12:34 crc kubenswrapper[4690]: I1203 13:12:34.901797 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gknq8" event={"ID":"c8c81207-0aa4-4489-a0e5-ed680f9436d2","Type":"ContainerDied","Data":"6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e"} Dec 03 13:12:34 crc kubenswrapper[4690]: I1203 13:12:34.905786 4690 generic.go:334] "Generic (PLEG): container finished" podID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerID="1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a" exitCode=0 Dec 03 13:12:34 crc kubenswrapper[4690]: I1203 13:12:34.905880 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvd94" event={"ID":"c06438b0-cd29-4387-96fd-257f01c8fd11","Type":"ContainerDied","Data":"1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a"} Dec 03 13:12:34 crc kubenswrapper[4690]: I1203 13:12:34.910764 4690 generic.go:334] "Generic (PLEG): container finished" podID="b710c0e8-81a2-427b-a964-d80b57047be3" containerID="8addbb8360584bdba6f6da19e7c945ff33c7443e2e3364d40ae9147528c52a1f" exitCode=0 Dec 03 13:12:34 crc kubenswrapper[4690]: I1203 13:12:34.910867 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27bwv" event={"ID":"b710c0e8-81a2-427b-a964-d80b57047be3","Type":"ContainerDied","Data":"8addbb8360584bdba6f6da19e7c945ff33c7443e2e3364d40ae9147528c52a1f"} Dec 03 13:12:35 crc kubenswrapper[4690]: I1203 13:12:35.918125 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gknq8" event={"ID":"c8c81207-0aa4-4489-a0e5-ed680f9436d2","Type":"ContainerStarted","Data":"95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0"} Dec 03 13:12:35 crc kubenswrapper[4690]: I1203 13:12:35.920262 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvd94" event={"ID":"c06438b0-cd29-4387-96fd-257f01c8fd11","Type":"ContainerStarted","Data":"3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62"} Dec 03 13:12:35 crc kubenswrapper[4690]: I1203 13:12:35.921582 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27bwv" event={"ID":"b710c0e8-81a2-427b-a964-d80b57047be3","Type":"ContainerStarted","Data":"f0770e162c9ecd4c9c8140464d81b39340a2a06e9e1e304ac955a946f80b4003"} Dec 03 13:12:35 crc kubenswrapper[4690]: I1203 13:12:35.924436 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrxpz" event={"ID":"61bb089c-4d85-45a7-ad6a-b16d95511a29","Type":"ContainerStarted","Data":"e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0"} Dec 03 13:12:35 crc kubenswrapper[4690]: I1203 13:12:35.939122 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gknq8" podStartSLOduration=3.413688386 podStartE2EDuration="58.939100338s" podCreationTimestamp="2025-12-03 13:11:37 +0000 UTC" firstStartedPulling="2025-12-03 13:11:40.025391705 +0000 UTC m=+146.006312138" lastFinishedPulling="2025-12-03 13:12:35.550803657 +0000 UTC m=+201.531724090" observedRunningTime="2025-12-03 13:12:35.938968535 +0000 UTC m=+201.919888978" watchObservedRunningTime="2025-12-03 13:12:35.939100338 +0000 UTC m=+201.920020771" Dec 03 13:12:35 crc kubenswrapper[4690]: I1203 13:12:35.960545 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rvd94" podStartSLOduration=3.818182879 podStartE2EDuration="56.960518914s" podCreationTimestamp="2025-12-03 13:11:39 +0000 UTC" firstStartedPulling="2025-12-03 13:11:42.231251601 +0000 UTC m=+148.212172034" lastFinishedPulling="2025-12-03 13:12:35.373587636 +0000 UTC m=+201.354508069" observedRunningTime="2025-12-03 13:12:35.954446225 +0000 UTC m=+201.935366658" watchObservedRunningTime="2025-12-03 13:12:35.960518914 +0000 UTC m=+201.941439357" Dec 03 13:12:35 crc kubenswrapper[4690]: I1203 13:12:35.985275 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-27bwv" podStartSLOduration=2.945283776 podStartE2EDuration="55.985252471s" podCreationTimestamp="2025-12-03 13:11:40 +0000 UTC" firstStartedPulling="2025-12-03 13:11:42.276259953 +0000 UTC m=+148.257180386" lastFinishedPulling="2025-12-03 13:12:35.316228648 +0000 UTC m=+201.297149081" observedRunningTime="2025-12-03 13:12:35.976772803 +0000 UTC m=+201.957693236" watchObservedRunningTime="2025-12-03 13:12:35.985252471 +0000 UTC m=+201.966172924" Dec 03 13:12:35 crc kubenswrapper[4690]: I1203 13:12:35.996649 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hrxpz" podStartSLOduration=4.820746676 podStartE2EDuration="58.99663125s" podCreationTimestamp="2025-12-03 13:11:37 +0000 UTC" firstStartedPulling="2025-12-03 13:11:41.174858302 +0000 UTC m=+147.155778735" lastFinishedPulling="2025-12-03 13:12:35.350742876 +0000 UTC m=+201.331663309" observedRunningTime="2025-12-03 13:12:35.995077512 +0000 UTC m=+201.975997945" watchObservedRunningTime="2025-12-03 13:12:35.99663125 +0000 UTC m=+201.977551683" Dec 03 13:12:36 crc kubenswrapper[4690]: I1203 13:12:36.931896 4690 generic.go:334] "Generic (PLEG): container finished" podID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerID="f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7" exitCode=0 Dec 03 13:12:36 crc kubenswrapper[4690]: I1203 13:12:36.932109 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j9q8" event={"ID":"9227b45f-65e8-4ed0-877a-fa41f765078f","Type":"ContainerDied","Data":"f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7"} Dec 03 13:12:37 crc kubenswrapper[4690]: I1203 13:12:37.791577 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:12:37 crc kubenswrapper[4690]: I1203 13:12:37.791959 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:12:37 crc kubenswrapper[4690]: I1203 13:12:37.939236 4690 generic.go:334] "Generic (PLEG): container finished" podID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerID="5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8" exitCode=0 Dec 03 13:12:37 crc kubenswrapper[4690]: I1203 13:12:37.939272 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6pw9" event={"ID":"0907470e-4498-4ff5-b6e1-e4be228e581e","Type":"ContainerDied","Data":"5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8"} Dec 03 13:12:37 crc kubenswrapper[4690]: I1203 13:12:37.944782 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j9q8" event={"ID":"9227b45f-65e8-4ed0-877a-fa41f765078f","Type":"ContainerStarted","Data":"dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6"} Dec 03 13:12:37 crc kubenswrapper[4690]: I1203 13:12:37.981085 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5j9q8" podStartSLOduration=2.948680701 podStartE2EDuration="57.981064062s" podCreationTimestamp="2025-12-03 13:11:40 +0000 UTC" firstStartedPulling="2025-12-03 13:11:42.297258921 +0000 UTC m=+148.278179354" lastFinishedPulling="2025-12-03 13:12:37.329642282 +0000 UTC m=+203.310562715" observedRunningTime="2025-12-03 13:12:37.976935421 +0000 UTC m=+203.957855864" watchObservedRunningTime="2025-12-03 13:12:37.981064062 +0000 UTC m=+203.961984505" Dec 03 13:12:38 crc kubenswrapper[4690]: I1203 13:12:38.026229 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:12:38 crc kubenswrapper[4690]: I1203 13:12:38.388964 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:12:38 crc kubenswrapper[4690]: I1203 13:12:38.389014 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:12:38 crc kubenswrapper[4690]: I1203 13:12:38.456945 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.215252 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.215699 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.419325 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.586407 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.586449 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.627026 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.815905 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.816217 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.998173 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:12:40 crc kubenswrapper[4690]: I1203 13:12:40.998824 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:12:41 crc kubenswrapper[4690]: I1203 13:12:41.853771 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5j9q8" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="registry-server" probeResult="failure" output=< Dec 03 13:12:41 crc kubenswrapper[4690]: timeout: failed to connect service ":50051" within 1s Dec 03 13:12:41 crc kubenswrapper[4690]: > Dec 03 13:12:43 crc kubenswrapper[4690]: I1203 13:12:43.977854 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6pw9" event={"ID":"0907470e-4498-4ff5-b6e1-e4be228e581e","Type":"ContainerStarted","Data":"e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537"} Dec 03 13:12:44 crc kubenswrapper[4690]: I1203 13:12:44.985936 4690 generic.go:334] "Generic (PLEG): container finished" podID="48f93164-1544-42e2-82e0-60a681752349" containerID="64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf" exitCode=0 Dec 03 13:12:44 crc kubenswrapper[4690]: I1203 13:12:44.986052 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8wnd" event={"ID":"48f93164-1544-42e2-82e0-60a681752349","Type":"ContainerDied","Data":"64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf"} Dec 03 13:12:45 crc kubenswrapper[4690]: I1203 13:12:45.023712 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n6pw9" podStartSLOduration=6.062191108 podStartE2EDuration="1m8.023691575s" podCreationTimestamp="2025-12-03 13:11:37 +0000 UTC" firstStartedPulling="2025-12-03 13:11:41.179845302 +0000 UTC m=+147.160765735" lastFinishedPulling="2025-12-03 13:12:43.141345769 +0000 UTC m=+209.122266202" observedRunningTime="2025-12-03 13:12:45.020954687 +0000 UTC m=+211.001875120" watchObservedRunningTime="2025-12-03 13:12:45.023691575 +0000 UTC m=+211.004612008" Dec 03 13:12:45 crc kubenswrapper[4690]: I1203 13:12:45.346375 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-27bwv"] Dec 03 13:12:45 crc kubenswrapper[4690]: I1203 13:12:45.346671 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-27bwv" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" containerName="registry-server" containerID="cri-o://f0770e162c9ecd4c9c8140464d81b39340a2a06e9e1e304ac955a946f80b4003" gracePeriod=2 Dec 03 13:12:46 crc kubenswrapper[4690]: I1203 13:12:46.823975 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:12:46 crc kubenswrapper[4690]: I1203 13:12:46.824055 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:12:46 crc kubenswrapper[4690]: I1203 13:12:46.824117 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:12:46 crc kubenswrapper[4690]: I1203 13:12:46.824783 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:12:46 crc kubenswrapper[4690]: I1203 13:12:46.824931 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed" gracePeriod=600 Dec 03 13:12:47 crc kubenswrapper[4690]: I1203 13:12:47.840528 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.005361 4690 generic.go:334] "Generic (PLEG): container finished" podID="b710c0e8-81a2-427b-a964-d80b57047be3" containerID="f0770e162c9ecd4c9c8140464d81b39340a2a06e9e1e304ac955a946f80b4003" exitCode=0 Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.005466 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27bwv" event={"ID":"b710c0e8-81a2-427b-a964-d80b57047be3","Type":"ContainerDied","Data":"f0770e162c9ecd4c9c8140464d81b39340a2a06e9e1e304ac955a946f80b4003"} Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.007348 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed" exitCode=0 Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.007395 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed"} Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.256536 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.309304 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.309393 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.356229 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.359454 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-utilities\") pod \"b710c0e8-81a2-427b-a964-d80b57047be3\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.359603 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfdtz\" (UniqueName: \"kubernetes.io/projected/b710c0e8-81a2-427b-a964-d80b57047be3-kube-api-access-qfdtz\") pod \"b710c0e8-81a2-427b-a964-d80b57047be3\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.359631 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-catalog-content\") pod \"b710c0e8-81a2-427b-a964-d80b57047be3\" (UID: \"b710c0e8-81a2-427b-a964-d80b57047be3\") " Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.360740 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-utilities" (OuterVolumeSpecName: "utilities") pod "b710c0e8-81a2-427b-a964-d80b57047be3" (UID: "b710c0e8-81a2-427b-a964-d80b57047be3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.368335 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b710c0e8-81a2-427b-a964-d80b57047be3-kube-api-access-qfdtz" (OuterVolumeSpecName: "kube-api-access-qfdtz") pod "b710c0e8-81a2-427b-a964-d80b57047be3" (UID: "b710c0e8-81a2-427b-a964-d80b57047be3"). InnerVolumeSpecName "kube-api-access-qfdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.386561 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b710c0e8-81a2-427b-a964-d80b57047be3" (UID: "b710c0e8-81a2-427b-a964-d80b57047be3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.434953 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.461142 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfdtz\" (UniqueName: \"kubernetes.io/projected/b710c0e8-81a2-427b-a964-d80b57047be3-kube-api-access-qfdtz\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.461188 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:48 crc kubenswrapper[4690]: I1203 13:12:48.461199 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710c0e8-81a2-427b-a964-d80b57047be3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.020921 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27bwv" Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.020912 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27bwv" event={"ID":"b710c0e8-81a2-427b-a964-d80b57047be3","Type":"ContainerDied","Data":"6cc66ad87fed1f49b80d3281f2bc9cc530e7ca70161d04e6af1ff1f9e834a876"} Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.021065 4690 scope.go:117] "RemoveContainer" containerID="f0770e162c9ecd4c9c8140464d81b39340a2a06e9e1e304ac955a946f80b4003" Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.025373 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8wnd" event={"ID":"48f93164-1544-42e2-82e0-60a681752349","Type":"ContainerStarted","Data":"5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f"} Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.032639 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"c9e92c52ce940ca7b19ac9c6c34c60bff5ccafb59832fe7eb80f99d62ea7a9e8"} Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.036672 4690 generic.go:334] "Generic (PLEG): container finished" podID="6a9984da-231b-45bf-8583-6394924d2e3e" containerID="7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e" exitCode=0 Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.036745 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drrqc" event={"ID":"6a9984da-231b-45bf-8583-6394924d2e3e","Type":"ContainerDied","Data":"7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e"} Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.047847 4690 scope.go:117] "RemoveContainer" containerID="8addbb8360584bdba6f6da19e7c945ff33c7443e2e3364d40ae9147528c52a1f" Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.059550 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f8wnd" podStartSLOduration=4.101088321 podStartE2EDuration="1m12.05952531s" podCreationTimestamp="2025-12-03 13:11:37 +0000 UTC" firstStartedPulling="2025-12-03 13:11:39.937031828 +0000 UTC m=+145.917952261" lastFinishedPulling="2025-12-03 13:12:47.895468817 +0000 UTC m=+213.876389250" observedRunningTime="2025-12-03 13:12:49.048591522 +0000 UTC m=+215.029511955" watchObservedRunningTime="2025-12-03 13:12:49.05952531 +0000 UTC m=+215.040445743" Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.140094 4690 scope.go:117] "RemoveContainer" containerID="62a63967f4fbf8cdf5e267efef15b4da28f6c3917077c6b41ce51b3547d7ba70" Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.142307 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.158343 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-27bwv"] Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.161122 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-27bwv"] Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.546979 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrxpz"] Dec 03 13:12:49 crc kubenswrapper[4690]: I1203 13:12:49.547244 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hrxpz" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerName="registry-server" containerID="cri-o://e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0" gracePeriod=2 Dec 03 13:12:50 crc kubenswrapper[4690]: I1203 13:12:50.321016 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" path="/var/lib/kubelet/pods/b710c0e8-81a2-427b-a964-d80b57047be3/volumes" Dec 03 13:12:50 crc kubenswrapper[4690]: I1203 13:12:50.861637 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:12:50 crc kubenswrapper[4690]: I1203 13:12:50.903004 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:12:51 crc kubenswrapper[4690]: I1203 13:12:51.947273 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6pw9"] Dec 03 13:12:51 crc kubenswrapper[4690]: I1203 13:12:51.947791 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n6pw9" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerName="registry-server" containerID="cri-o://e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537" gracePeriod=2 Dec 03 13:12:51 crc kubenswrapper[4690]: I1203 13:12:51.957369 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.005516 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-catalog-content\") pod \"61bb089c-4d85-45a7-ad6a-b16d95511a29\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.005596 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwsrl\" (UniqueName: \"kubernetes.io/projected/61bb089c-4d85-45a7-ad6a-b16d95511a29-kube-api-access-zwsrl\") pod \"61bb089c-4d85-45a7-ad6a-b16d95511a29\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.005675 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-utilities\") pod \"61bb089c-4d85-45a7-ad6a-b16d95511a29\" (UID: \"61bb089c-4d85-45a7-ad6a-b16d95511a29\") " Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.006995 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-utilities" (OuterVolumeSpecName: "utilities") pod "61bb089c-4d85-45a7-ad6a-b16d95511a29" (UID: "61bb089c-4d85-45a7-ad6a-b16d95511a29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.012570 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61bb089c-4d85-45a7-ad6a-b16d95511a29-kube-api-access-zwsrl" (OuterVolumeSpecName: "kube-api-access-zwsrl") pod "61bb089c-4d85-45a7-ad6a-b16d95511a29" (UID: "61bb089c-4d85-45a7-ad6a-b16d95511a29"). InnerVolumeSpecName "kube-api-access-zwsrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.055396 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61bb089c-4d85-45a7-ad6a-b16d95511a29" (UID: "61bb089c-4d85-45a7-ad6a-b16d95511a29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.055760 4690 generic.go:334] "Generic (PLEG): container finished" podID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerID="e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0" exitCode=0 Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.055814 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrxpz" event={"ID":"61bb089c-4d85-45a7-ad6a-b16d95511a29","Type":"ContainerDied","Data":"e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0"} Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.056051 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrxpz" event={"ID":"61bb089c-4d85-45a7-ad6a-b16d95511a29","Type":"ContainerDied","Data":"9e2f9e5fb7b1e947564d0a375ea876a6d437ec0dc8cb1b3326b36e7fe753ac36"} Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.055829 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrxpz" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.056082 4690 scope.go:117] "RemoveContainer" containerID="e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.087060 4690 scope.go:117] "RemoveContainer" containerID="412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.099891 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrxpz"] Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.103916 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hrxpz"] Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.107112 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.107241 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61bb089c-4d85-45a7-ad6a-b16d95511a29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.107324 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwsrl\" (UniqueName: \"kubernetes.io/projected/61bb089c-4d85-45a7-ad6a-b16d95511a29-kube-api-access-zwsrl\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.117897 4690 scope.go:117] "RemoveContainer" containerID="3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.131300 4690 scope.go:117] "RemoveContainer" containerID="e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0" Dec 03 13:12:52 crc kubenswrapper[4690]: E1203 13:12:52.131684 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0\": container with ID starting with e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0 not found: ID does not exist" containerID="e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.131719 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0"} err="failed to get container status \"e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0\": rpc error: code = NotFound desc = could not find container \"e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0\": container with ID starting with e7a5718459b8870c7ba4b224d08daddf0dad6e8d89fea679e3cd47bf7657d6d0 not found: ID does not exist" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.131745 4690 scope.go:117] "RemoveContainer" containerID="412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8" Dec 03 13:12:52 crc kubenswrapper[4690]: E1203 13:12:52.132146 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8\": container with ID starting with 412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8 not found: ID does not exist" containerID="412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.132209 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8"} err="failed to get container status \"412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8\": rpc error: code = NotFound desc = could not find container \"412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8\": container with ID starting with 412a0cf847c9ed6bd4e19db6fe0762629b1e0cc1b41c9a68b775c587148ddea8 not found: ID does not exist" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.132249 4690 scope.go:117] "RemoveContainer" containerID="3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733" Dec 03 13:12:52 crc kubenswrapper[4690]: E1203 13:12:52.132741 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733\": container with ID starting with 3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733 not found: ID does not exist" containerID="3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.132772 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733"} err="failed to get container status \"3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733\": rpc error: code = NotFound desc = could not find container \"3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733\": container with ID starting with 3f2b27b4a6078631b50b8b6bf63f39018e1257df47afd5313972940d4f5fb733 not found: ID does not exist" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.267238 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.309631 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjbsh\" (UniqueName: \"kubernetes.io/projected/0907470e-4498-4ff5-b6e1-e4be228e581e-kube-api-access-tjbsh\") pod \"0907470e-4498-4ff5-b6e1-e4be228e581e\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.309709 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-utilities\") pod \"0907470e-4498-4ff5-b6e1-e4be228e581e\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.309764 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-catalog-content\") pod \"0907470e-4498-4ff5-b6e1-e4be228e581e\" (UID: \"0907470e-4498-4ff5-b6e1-e4be228e581e\") " Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.310792 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-utilities" (OuterVolumeSpecName: "utilities") pod "0907470e-4498-4ff5-b6e1-e4be228e581e" (UID: "0907470e-4498-4ff5-b6e1-e4be228e581e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.312403 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0907470e-4498-4ff5-b6e1-e4be228e581e-kube-api-access-tjbsh" (OuterVolumeSpecName: "kube-api-access-tjbsh") pod "0907470e-4498-4ff5-b6e1-e4be228e581e" (UID: "0907470e-4498-4ff5-b6e1-e4be228e581e"). InnerVolumeSpecName "kube-api-access-tjbsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.323109 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" path="/var/lib/kubelet/pods/61bb089c-4d85-45a7-ad6a-b16d95511a29/volumes" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.358730 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0907470e-4498-4ff5-b6e1-e4be228e581e" (UID: "0907470e-4498-4ff5-b6e1-e4be228e581e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.411314 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjbsh\" (UniqueName: \"kubernetes.io/projected/0907470e-4498-4ff5-b6e1-e4be228e581e-kube-api-access-tjbsh\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.411367 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:52 crc kubenswrapper[4690]: I1203 13:12:52.411380 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0907470e-4498-4ff5-b6e1-e4be228e581e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.062963 4690 generic.go:334] "Generic (PLEG): container finished" podID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerID="e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537" exitCode=0 Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.063007 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6pw9" event={"ID":"0907470e-4498-4ff5-b6e1-e4be228e581e","Type":"ContainerDied","Data":"e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537"} Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.063031 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6pw9" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.063040 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6pw9" event={"ID":"0907470e-4498-4ff5-b6e1-e4be228e581e","Type":"ContainerDied","Data":"740aed24885ddc7375a867daf36857c3fe8f074a0eea4fb2b0af6ca0e07dec14"} Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.063062 4690 scope.go:117] "RemoveContainer" containerID="e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.087366 4690 scope.go:117] "RemoveContainer" containerID="5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.092764 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6pw9"] Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.095156 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n6pw9"] Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.119947 4690 scope.go:117] "RemoveContainer" containerID="b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.133192 4690 scope.go:117] "RemoveContainer" containerID="e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537" Dec 03 13:12:53 crc kubenswrapper[4690]: E1203 13:12:53.133762 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537\": container with ID starting with e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537 not found: ID does not exist" containerID="e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.133813 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537"} err="failed to get container status \"e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537\": rpc error: code = NotFound desc = could not find container \"e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537\": container with ID starting with e0f5c51dddc9758dcef4a8d08c1fa22cb991ada754ab6f09068eb824e0065537 not found: ID does not exist" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.133854 4690 scope.go:117] "RemoveContainer" containerID="5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8" Dec 03 13:12:53 crc kubenswrapper[4690]: E1203 13:12:53.134130 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8\": container with ID starting with 5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8 not found: ID does not exist" containerID="5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.134161 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8"} err="failed to get container status \"5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8\": rpc error: code = NotFound desc = could not find container \"5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8\": container with ID starting with 5c641e537081d7754da2e02a539c038cbc9a23407095271489117ab0f74714d8 not found: ID does not exist" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.134180 4690 scope.go:117] "RemoveContainer" containerID="b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4" Dec 03 13:12:53 crc kubenswrapper[4690]: E1203 13:12:53.134383 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4\": container with ID starting with b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4 not found: ID does not exist" containerID="b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4" Dec 03 13:12:53 crc kubenswrapper[4690]: I1203 13:12:53.134412 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4"} err="failed to get container status \"b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4\": rpc error: code = NotFound desc = could not find container \"b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4\": container with ID starting with b4e2e616061aec349bec602fbf8adf30a9b535f3fa34d2087089ef1547a561f4 not found: ID does not exist" Dec 03 13:12:54 crc kubenswrapper[4690]: I1203 13:12:54.323180 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" path="/var/lib/kubelet/pods/0907470e-4498-4ff5-b6e1-e4be228e581e/volumes" Dec 03 13:12:54 crc kubenswrapper[4690]: I1203 13:12:54.950025 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" podUID="d5c2b268-0de1-4c88-b439-5292c56044dd" containerName="oauth-openshift" containerID="cri-o://0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304" gracePeriod=15 Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.801905 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.854658 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-ocp-branding-template\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.854757 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-error\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.854835 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-dir\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.854942 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-session\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.854995 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-login\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855044 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-idp-0-file-data\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855155 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-provider-selection\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855210 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-router-certs\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855270 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-trusted-ca-bundle\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855338 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4xxl\" (UniqueName: \"kubernetes.io/projected/d5c2b268-0de1-4c88-b439-5292c56044dd-kube-api-access-x4xxl\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855391 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-cliconfig\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855459 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-policies\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855495 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-serving-cert\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.855545 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-service-ca\") pod \"d5c2b268-0de1-4c88-b439-5292c56044dd\" (UID: \"d5c2b268-0de1-4c88-b439-5292c56044dd\") " Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.856557 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.857889 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.857921 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.858415 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.858928 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.862941 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.863698 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.863721 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c2b268-0de1-4c88-b439-5292c56044dd-kube-api-access-x4xxl" (OuterVolumeSpecName: "kube-api-access-x4xxl") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "kube-api-access-x4xxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.864122 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.864453 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.864937 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.864996 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.866841 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.875249 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "d5c2b268-0de1-4c88-b439-5292c56044dd" (UID: "d5c2b268-0de1-4c88-b439-5292c56044dd"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.957154 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.957821 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.957920 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958023 4690 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958129 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958222 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958292 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958391 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958500 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958583 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958663 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4xxl\" (UniqueName: \"kubernetes.io/projected/d5c2b268-0de1-4c88-b439-5292c56044dd-kube-api-access-x4xxl\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958746 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958824 4690 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d5c2b268-0de1-4c88-b439-5292c56044dd-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:55 crc kubenswrapper[4690]: I1203 13:12:55.958966 4690 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5c2b268-0de1-4c88-b439-5292c56044dd-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.086907 4690 generic.go:334] "Generic (PLEG): container finished" podID="d5c2b268-0de1-4c88-b439-5292c56044dd" containerID="0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304" exitCode=0 Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.086971 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" event={"ID":"d5c2b268-0de1-4c88-b439-5292c56044dd","Type":"ContainerDied","Data":"0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304"} Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.086994 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.087018 4690 scope.go:117] "RemoveContainer" containerID="0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.087003 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-54f5g" event={"ID":"d5c2b268-0de1-4c88-b439-5292c56044dd","Type":"ContainerDied","Data":"70f4c4da7dce26de4d5962dce8f7781ffeecb4ddbef91635d525650ef02f0e21"} Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.111679 4690 scope.go:117] "RemoveContainer" containerID="0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.115189 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304\": container with ID starting with 0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304 not found: ID does not exist" containerID="0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.115260 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304"} err="failed to get container status \"0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304\": rpc error: code = NotFound desc = could not find container \"0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304\": container with ID starting with 0bf6a1d0ceecca0af1f8b876d97ee7c96083213c4d9a54bbb3202313c704b304 not found: ID does not exist" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.123814 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54f5g"] Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.127144 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54f5g"] Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.322247 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5c2b268-0de1-4c88-b439-5292c56044dd" path="/var/lib/kubelet/pods/d5c2b268-0de1-4c88-b439-5292c56044dd/volumes" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396222 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7"] Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396488 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerName="extract-content" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396505 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerName="extract-content" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396519 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerName="extract-content" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396526 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerName="extract-content" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396546 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" containerName="extract-utilities" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396554 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" containerName="extract-utilities" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396566 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396573 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396587 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" containerName="extract-content" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396594 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" containerName="extract-content" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396605 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerName="extract-utilities" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396613 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerName="extract-utilities" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396625 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396633 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396641 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c2b268-0de1-4c88-b439-5292c56044dd" containerName="oauth-openshift" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396648 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c2b268-0de1-4c88-b439-5292c56044dd" containerName="oauth-openshift" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396659 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396666 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: E1203 13:12:56.396678 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerName="extract-utilities" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396686 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerName="extract-utilities" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396819 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bb089c-4d85-45a7-ad6a-b16d95511a29" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396848 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="0907470e-4498-4ff5-b6e1-e4be228e581e" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396877 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c2b268-0de1-4c88-b439-5292c56044dd" containerName="oauth-openshift" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.396888 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b710c0e8-81a2-427b-a964-d80b57047be3" containerName="registry-server" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.397469 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.401190 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.401283 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.401317 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.401388 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.401514 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.401546 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.401583 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.401817 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.403330 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.403648 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.404121 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.406051 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.410356 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7"] Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.413039 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.416804 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.433359 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465065 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465140 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz5v5\" (UniqueName: \"kubernetes.io/projected/204f76a9-359e-4a47-833c-14191e13defa-kube-api-access-zz5v5\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465174 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465207 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465241 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465334 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/204f76a9-359e-4a47-833c-14191e13defa-audit-dir\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465422 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465499 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-audit-policies\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465549 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-error\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465584 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-login\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465616 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465640 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465655 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.465714 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-session\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567179 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-session\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567238 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567274 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz5v5\" (UniqueName: \"kubernetes.io/projected/204f76a9-359e-4a47-833c-14191e13defa-kube-api-access-zz5v5\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567299 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567327 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567356 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567382 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/204f76a9-359e-4a47-833c-14191e13defa-audit-dir\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567412 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567450 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-audit-policies\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567478 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-error\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567504 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-login\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567528 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567550 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.567569 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.568098 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/204f76a9-359e-4a47-833c-14191e13defa-audit-dir\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.568297 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-audit-policies\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.568691 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.568949 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.569334 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.571710 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.571923 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.572086 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-error\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.572818 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.576251 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.581241 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-user-template-login\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.582828 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz5v5\" (UniqueName: \"kubernetes.io/projected/204f76a9-359e-4a47-833c-14191e13defa-kube-api-access-zz5v5\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.583227 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-session\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.584417 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/204f76a9-359e-4a47-833c-14191e13defa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cf78f7d5b-wt6f7\" (UID: \"204f76a9-359e-4a47-833c-14191e13defa\") " pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:56 crc kubenswrapper[4690]: I1203 13:12:56.822477 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:57 crc kubenswrapper[4690]: I1203 13:12:57.272266 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7"] Dec 03 13:12:58 crc kubenswrapper[4690]: I1203 13:12:58.103785 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" event={"ID":"204f76a9-359e-4a47-833c-14191e13defa","Type":"ContainerStarted","Data":"f825501345949df7caf38322be138fc372ae308a1047e2622ec9d4340ea42b08"} Dec 03 13:12:58 crc kubenswrapper[4690]: I1203 13:12:58.104188 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" event={"ID":"204f76a9-359e-4a47-833c-14191e13defa","Type":"ContainerStarted","Data":"bd164f1fcb4dc54fb8769472ed423a23f5dcb526b40a0f77b62e7be3edc08354"} Dec 03 13:12:58 crc kubenswrapper[4690]: I1203 13:12:58.104277 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:58 crc kubenswrapper[4690]: I1203 13:12:58.112353 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" Dec 03 13:12:58 crc kubenswrapper[4690]: I1203 13:12:58.127593 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7cf78f7d5b-wt6f7" podStartSLOduration=29.127566401 podStartE2EDuration="29.127566401s" podCreationTimestamp="2025-12-03 13:12:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:12:58.125131311 +0000 UTC m=+224.106051744" watchObservedRunningTime="2025-12-03 13:12:58.127566401 +0000 UTC m=+224.108486844" Dec 03 13:12:58 crc kubenswrapper[4690]: I1203 13:12:58.179439 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:12:58 crc kubenswrapper[4690]: I1203 13:12:58.179786 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:12:58 crc kubenswrapper[4690]: I1203 13:12:58.237372 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:12:59 crc kubenswrapper[4690]: I1203 13:12:59.156894 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:12:59 crc kubenswrapper[4690]: E1203 13:12:59.499402 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/b8/b856e4d37af238240aaa3504ebf72881a05d3e5875365377d4fbd3a313fe7d06?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251203%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251203T131249Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=314b777610af42bd657dee6982c0f8a8a7716c9937fac7a3ebd26e31c9ed244a®ion=us-east-1&namespace=openshift-release-dev&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=ocp-v4.0-art-dev&akamai_signature=exp=1764768469~hmac=dc836671681c02d08736f6e3ab5834582cf43bb78bea76ee672bccad9a52bf90\": net/http: TLS handshake timeout" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad" Dec 03 13:12:59 crc kubenswrapper[4690]: E1203 13:12:59.499856 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad,Command:[/bin/opm],Args:[serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOMEMLIMIT,Value:30MiB,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{31457280 0} {} 30Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5g77k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-drrqc_openshift-marketplace(6a9984da-231b-45bf-8583-6394924d2e3e): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/b8/b856e4d37af238240aaa3504ebf72881a05d3e5875365377d4fbd3a313fe7d06?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251203%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251203T131249Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=314b777610af42bd657dee6982c0f8a8a7716c9937fac7a3ebd26e31c9ed244a®ion=us-east-1&namespace=openshift-release-dev&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=ocp-v4.0-art-dev&akamai_signature=exp=1764768469~hmac=dc836671681c02d08736f6e3ab5834582cf43bb78bea76ee672bccad9a52bf90\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 13:12:59 crc kubenswrapper[4690]: E1203 13:12:59.501227 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/b8/b856e4d37af238240aaa3504ebf72881a05d3e5875365377d4fbd3a313fe7d06?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251203%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251203T131249Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=314b777610af42bd657dee6982c0f8a8a7716c9937fac7a3ebd26e31c9ed244a®ion=us-east-1&namespace=openshift-release-dev&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=ocp-v4.0-art-dev&akamai_signature=exp=1764768469~hmac=dc836671681c02d08736f6e3ab5834582cf43bb78bea76ee672bccad9a52bf90\\\": net/http: TLS handshake timeout\"" pod="openshift-marketplace/redhat-operators-drrqc" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.285378 4690 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.286638 4690 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.286899 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984" gracePeriod=15 Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.287110 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.287347 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d" gracePeriod=15 Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.287461 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539" gracePeriod=15 Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.287445 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44" gracePeriod=15 Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.287606 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c" gracePeriod=15 Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.287891 4690 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:13:07 crc kubenswrapper[4690]: E1203 13:13:07.288023 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288041 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 13:13:07 crc kubenswrapper[4690]: E1203 13:13:07.288058 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288067 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 13:13:07 crc kubenswrapper[4690]: E1203 13:13:07.288080 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288088 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 13:13:07 crc kubenswrapper[4690]: E1203 13:13:07.288097 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288106 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 13:13:07 crc kubenswrapper[4690]: E1203 13:13:07.288120 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288128 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:13:07 crc kubenswrapper[4690]: E1203 13:13:07.288134 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288140 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 13:13:07 crc kubenswrapper[4690]: E1203 13:13:07.288147 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288154 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288262 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288270 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288279 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288287 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288297 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.288306 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.331565 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.413883 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.413948 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.413994 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.414094 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.414220 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.414303 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.414344 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.414379 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515312 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515385 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515418 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515439 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515457 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515483 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515500 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515528 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515570 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515609 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515678 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515746 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515754 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515794 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515842 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.515934 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: I1203 13:13:07.626567 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:13:07 crc kubenswrapper[4690]: W1203 13:13:07.658148 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-d6b971f59ef65beb4ebfeda01c4cf19a95f51d2e596eb3c85f21fd21dd05b34f WatchSource:0}: Error finding container d6b971f59ef65beb4ebfeda01c4cf19a95f51d2e596eb3c85f21fd21dd05b34f: Status 404 returned error can't find the container with id d6b971f59ef65beb4ebfeda01c4cf19a95f51d2e596eb3c85f21fd21dd05b34f Dec 03 13:13:08 crc kubenswrapper[4690]: I1203 13:13:08.160954 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d6b971f59ef65beb4ebfeda01c4cf19a95f51d2e596eb3c85f21fd21dd05b34f"} Dec 03 13:13:08 crc kubenswrapper[4690]: E1203 13:13:08.601313 4690 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187db6c58d9dbb14 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 13:13:08.599839508 +0000 UTC m=+234.580759941,LastTimestamp:2025-12-03 13:13:08.599839508 +0000 UTC m=+234.580759941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.169530 4690 generic.go:334] "Generic (PLEG): container finished" podID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" containerID="852a183a677ab4500648ed661080ab91234e747362104ae04d4cb1c22865308d" exitCode=0 Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.169930 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"455990f2-8fc3-4aa8-9a20-1c1f4d747d51","Type":"ContainerDied","Data":"852a183a677ab4500648ed661080ab91234e747362104ae04d4cb1c22865308d"} Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.170645 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.170854 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.172856 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.174121 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.174841 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c" exitCode=0 Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.174890 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44" exitCode=0 Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.174901 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d" exitCode=0 Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.174908 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539" exitCode=2 Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.174996 4690 scope.go:117] "RemoveContainer" containerID="85bf9a47b46a51906ca884d8df12d3ec297e6225681434839c583832fbfd4fff" Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.177895 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"bd2b6627fd9da4c9bcaddf8ea22923cbefe6a05bbd1559eb000aeebeae998a6f"} Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.181443 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.181859 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: E1203 13:13:09.799108 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: E1203 13:13:09.799937 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: E1203 13:13:09.800543 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: E1203 13:13:09.800984 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: E1203 13:13:09.801696 4690 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:09 crc kubenswrapper[4690]: I1203 13:13:09.801726 4690 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 13:13:09 crc kubenswrapper[4690]: E1203 13:13:09.801960 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="200ms" Dec 03 13:13:09 crc kubenswrapper[4690]: E1203 13:13:09.848487 4690 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187db6c58d9dbb14 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 13:13:08.599839508 +0000 UTC m=+234.580759941,LastTimestamp:2025-12-03 13:13:08.599839508 +0000 UTC m=+234.580759941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 13:13:10 crc kubenswrapper[4690]: E1203 13:13:10.003731 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="400ms" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.185127 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 13:13:10 crc kubenswrapper[4690]: E1203 13:13:10.405233 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="800ms" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.442536 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.443698 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.444243 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.554446 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kube-api-access\") pod \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.554565 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-var-lock\") pod \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.554667 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-var-lock" (OuterVolumeSpecName: "var-lock") pod "455990f2-8fc3-4aa8-9a20-1c1f4d747d51" (UID: "455990f2-8fc3-4aa8-9a20-1c1f4d747d51"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.554700 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kubelet-dir\") pod \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\" (UID: \"455990f2-8fc3-4aa8-9a20-1c1f4d747d51\") " Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.554819 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "455990f2-8fc3-4aa8-9a20-1c1f4d747d51" (UID: "455990f2-8fc3-4aa8-9a20-1c1f4d747d51"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.555209 4690 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.555272 4690 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.561929 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "455990f2-8fc3-4aa8-9a20-1c1f4d747d51" (UID: "455990f2-8fc3-4aa8-9a20-1c1f4d747d51"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.657769 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/455990f2-8fc3-4aa8-9a20-1c1f4d747d51-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.892969 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.894265 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.895221 4690 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.895803 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:10 crc kubenswrapper[4690]: I1203 13:13:10.896319 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.062160 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.062210 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.062250 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.062306 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.062315 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.062444 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.062470 4690 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.062533 4690 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.163439 4690 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.192255 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"455990f2-8fc3-4aa8-9a20-1c1f4d747d51","Type":"ContainerDied","Data":"8126d0a1726b25c140029123d5359015ce6deedb14ad3079cc5111c8ce7cf64d"} Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.192295 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.192318 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8126d0a1726b25c140029123d5359015ce6deedb14ad3079cc5111c8ce7cf64d" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.195057 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.195727 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984" exitCode=0 Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.195776 4690 scope.go:117] "RemoveContainer" containerID="4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.195955 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:11 crc kubenswrapper[4690]: E1203 13:13:11.206450 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="1.6s" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.209600 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.209789 4690 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.210186 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.212360 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.212577 4690 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.212743 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.215326 4690 scope.go:117] "RemoveContainer" containerID="70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.230914 4690 scope.go:117] "RemoveContainer" containerID="e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.245265 4690 scope.go:117] "RemoveContainer" containerID="5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.259441 4690 scope.go:117] "RemoveContainer" containerID="5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.275470 4690 scope.go:117] "RemoveContainer" containerID="647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.292705 4690 scope.go:117] "RemoveContainer" containerID="4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c" Dec 03 13:13:11 crc kubenswrapper[4690]: E1203 13:13:11.293219 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\": container with ID starting with 4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c not found: ID does not exist" containerID="4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.293282 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c"} err="failed to get container status \"4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\": rpc error: code = NotFound desc = could not find container \"4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c\": container with ID starting with 4140893678bb74f1e66c82c037abdcb94ab67d48bcb1b05688864e9a4063c18c not found: ID does not exist" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.293322 4690 scope.go:117] "RemoveContainer" containerID="70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44" Dec 03 13:13:11 crc kubenswrapper[4690]: E1203 13:13:11.293780 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\": container with ID starting with 70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44 not found: ID does not exist" containerID="70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.293813 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44"} err="failed to get container status \"70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\": rpc error: code = NotFound desc = could not find container \"70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44\": container with ID starting with 70c758332e380d6b9f3e3dfb379e379f45b98bbe61665e6f50dc48e734f0ef44 not found: ID does not exist" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.293836 4690 scope.go:117] "RemoveContainer" containerID="e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d" Dec 03 13:13:11 crc kubenswrapper[4690]: E1203 13:13:11.294131 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\": container with ID starting with e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d not found: ID does not exist" containerID="e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.294153 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d"} err="failed to get container status \"e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\": rpc error: code = NotFound desc = could not find container \"e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d\": container with ID starting with e1fea0892bfe6bf2f5914a33eff95311ff71d2cd8ae93a76632c0fceb366f33d not found: ID does not exist" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.294170 4690 scope.go:117] "RemoveContainer" containerID="5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539" Dec 03 13:13:11 crc kubenswrapper[4690]: E1203 13:13:11.294427 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\": container with ID starting with 5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539 not found: ID does not exist" containerID="5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.294451 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539"} err="failed to get container status \"5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\": rpc error: code = NotFound desc = could not find container \"5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539\": container with ID starting with 5286c0a6b95c49b81682ef48ca22bb4f99513479ad420d3335174d626fda9539 not found: ID does not exist" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.294494 4690 scope.go:117] "RemoveContainer" containerID="5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984" Dec 03 13:13:11 crc kubenswrapper[4690]: E1203 13:13:11.294733 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\": container with ID starting with 5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984 not found: ID does not exist" containerID="5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.294759 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984"} err="failed to get container status \"5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\": rpc error: code = NotFound desc = could not find container \"5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984\": container with ID starting with 5da5617468043dcc7b95b21d61ada9dcea3c46a885d7d8a24877054abba95984 not found: ID does not exist" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.294774 4690 scope.go:117] "RemoveContainer" containerID="647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f" Dec 03 13:13:11 crc kubenswrapper[4690]: E1203 13:13:11.295119 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\": container with ID starting with 647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f not found: ID does not exist" containerID="647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f" Dec 03 13:13:11 crc kubenswrapper[4690]: I1203 13:13:11.295145 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f"} err="failed to get container status \"647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\": rpc error: code = NotFound desc = could not find container \"647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f\": container with ID starting with 647df48f9a66e5b4634f3095d26dd9d3e253e4415d217cad83d3265c8b6dd05f not found: ID does not exist" Dec 03 13:13:12 crc kubenswrapper[4690]: I1203 13:13:12.320779 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 13:13:12 crc kubenswrapper[4690]: E1203 13:13:12.807160 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="3.2s" Dec 03 13:13:14 crc kubenswrapper[4690]: I1203 13:13:14.318356 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:14 crc kubenswrapper[4690]: I1203 13:13:14.319023 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:14 crc kubenswrapper[4690]: I1203 13:13:14.319421 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:14 crc kubenswrapper[4690]: I1203 13:13:14.320511 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:14 crc kubenswrapper[4690]: I1203 13:13:14.322301 4690 status_manager.go:851] "Failed to get status for pod" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" pod="openshift-marketplace/redhat-operators-drrqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-drrqc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:15 crc kubenswrapper[4690]: I1203 13:13:15.225897 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drrqc" event={"ID":"6a9984da-231b-45bf-8583-6394924d2e3e","Type":"ContainerStarted","Data":"ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68"} Dec 03 13:13:15 crc kubenswrapper[4690]: I1203 13:13:15.228524 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:15 crc kubenswrapper[4690]: I1203 13:13:15.228916 4690 status_manager.go:851] "Failed to get status for pod" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" pod="openshift-marketplace/redhat-operators-drrqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-drrqc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:15 crc kubenswrapper[4690]: I1203 13:13:15.229408 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:16 crc kubenswrapper[4690]: E1203 13:13:16.008525 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="6.4s" Dec 03 13:13:19 crc kubenswrapper[4690]: E1203 13:13:19.849824 4690 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187db6c58d9dbb14 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 13:13:08.599839508 +0000 UTC m=+234.580759941,LastTimestamp:2025-12-03 13:13:08.599839508 +0000 UTC m=+234.580759941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.092098 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.092943 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.131852 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.132628 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.133182 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.133625 4690 status_manager.go:851] "Failed to get status for pod" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" pod="openshift-marketplace/redhat-operators-drrqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-drrqc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.290246 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.290852 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.291316 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:21 crc kubenswrapper[4690]: I1203 13:13:21.291668 4690 status_manager.go:851] "Failed to get status for pod" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" pod="openshift-marketplace/redhat-operators-drrqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-drrqc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.264335 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.264695 4690 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1" exitCode=1 Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.264962 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1"} Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.265649 4690 scope.go:117] "RemoveContainer" containerID="fd912e55bfeb7ae08268dd32b01dd3cc679d5851557b161b0eade4a8fffca5e1" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.265929 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.266519 4690 status_manager.go:851] "Failed to get status for pod" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" pod="openshift-marketplace/redhat-operators-drrqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-drrqc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.266741 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.267024 4690 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.314295 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.315412 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.315969 4690 status_manager.go:851] "Failed to get status for pod" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" pod="openshift-marketplace/redhat-operators-drrqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-drrqc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.316549 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.317078 4690 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.346676 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.346710 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:22 crc kubenswrapper[4690]: E1203 13:13:22.347175 4690 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:22 crc kubenswrapper[4690]: I1203 13:13:22.347833 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:22 crc kubenswrapper[4690]: E1203 13:13:22.409802 4690 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="7s" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.272592 4690 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="c608a4e0d34a35994302096f21ecb98d85c87407b793275185c37694ff53ee69" exitCode=0 Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.272740 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"c608a4e0d34a35994302096f21ecb98d85c87407b793275185c37694ff53ee69"} Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.273041 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"63d15d7219513df35009c13d6ea53412e8481cfc63f9309991f8e46a5a88b9ec"} Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.273328 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.273340 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:23 crc kubenswrapper[4690]: E1203 13:13:23.273751 4690 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.273816 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.274141 4690 status_manager.go:851] "Failed to get status for pod" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" pod="openshift-marketplace/redhat-operators-drrqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-drrqc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.274461 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.274703 4690 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.277009 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.277312 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7987f2a5e0cab9ed7b2c784ef8c1ba55be085102ab238da453a9a771decd857b"} Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.277785 4690 status_manager.go:851] "Failed to get status for pod" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.278155 4690 status_manager.go:851] "Failed to get status for pod" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" pod="openshift-marketplace/redhat-operators-drrqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-drrqc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.278632 4690 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:23 crc kubenswrapper[4690]: I1203 13:13:23.279112 4690 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 03 13:13:24 crc kubenswrapper[4690]: I1203 13:13:24.294627 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ee1df3d016afaa8a5a595cc333c26ef42fee7439fbf392c8b32197b6808f3dbe"} Dec 03 13:13:24 crc kubenswrapper[4690]: I1203 13:13:24.295072 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c0ec4dd1e6e28f542a290c1e24dc03041ea95052b8769797d898501b3fa9cd25"} Dec 03 13:13:24 crc kubenswrapper[4690]: I1203 13:13:24.295088 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d8710c7e5c1842eb688539d1ea7ed70c429dd58f8232148654eb209694e2cb8c"} Dec 03 13:13:24 crc kubenswrapper[4690]: I1203 13:13:24.295099 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d6789fe02d565c260751ecff9b35117a64c6505799f7313e92d83dca955505d5"} Dec 03 13:13:25 crc kubenswrapper[4690]: I1203 13:13:25.302556 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b314c8ab0b9a30429c3e49e6d04dbdb1c5cb3efb50d44bf407aa178af6692b0a"} Dec 03 13:13:25 crc kubenswrapper[4690]: I1203 13:13:25.302813 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:25 crc kubenswrapper[4690]: I1203 13:13:25.302926 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:25 crc kubenswrapper[4690]: I1203 13:13:25.302951 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:27 crc kubenswrapper[4690]: I1203 13:13:27.348304 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:27 crc kubenswrapper[4690]: I1203 13:13:27.348655 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:27 crc kubenswrapper[4690]: I1203 13:13:27.355534 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:29 crc kubenswrapper[4690]: I1203 13:13:29.242447 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:13:29 crc kubenswrapper[4690]: I1203 13:13:29.248847 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:13:29 crc kubenswrapper[4690]: I1203 13:13:29.322796 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:13:30 crc kubenswrapper[4690]: I1203 13:13:30.321187 4690 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:31 crc kubenswrapper[4690]: I1203 13:13:31.330925 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:31 crc kubenswrapper[4690]: I1203 13:13:31.330955 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:31 crc kubenswrapper[4690]: I1203 13:13:31.334671 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:13:31 crc kubenswrapper[4690]: I1203 13:13:31.337454 4690 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="e3c3c2d5-5fee-4d3e-93a4-9fb97df4ce5b" Dec 03 13:13:32 crc kubenswrapper[4690]: I1203 13:13:32.335013 4690 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:32 crc kubenswrapper[4690]: I1203 13:13:32.335397 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="89e2e662-ec6e-499e-a68f-b57d17aa218e" Dec 03 13:13:34 crc kubenswrapper[4690]: I1203 13:13:34.332922 4690 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="e3c3c2d5-5fee-4d3e-93a4-9fb97df4ce5b" Dec 03 13:13:39 crc kubenswrapper[4690]: I1203 13:13:39.749614 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:13:40 crc kubenswrapper[4690]: I1203 13:13:40.557576 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 13:13:40 crc kubenswrapper[4690]: I1203 13:13:40.693044 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 13:13:40 crc kubenswrapper[4690]: I1203 13:13:40.758885 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 13:13:40 crc kubenswrapper[4690]: I1203 13:13:40.949761 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 13:13:40 crc kubenswrapper[4690]: I1203 13:13:40.959360 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 13:13:40 crc kubenswrapper[4690]: I1203 13:13:40.979554 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 13:13:41 crc kubenswrapper[4690]: I1203 13:13:41.210313 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 13:13:41 crc kubenswrapper[4690]: I1203 13:13:41.384390 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 13:13:41 crc kubenswrapper[4690]: I1203 13:13:41.693738 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 13:13:41 crc kubenswrapper[4690]: I1203 13:13:41.829389 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.030581 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.321464 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.371845 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.519638 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.541369 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.573165 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.711907 4690 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.871583 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 13:13:42 crc kubenswrapper[4690]: I1203 13:13:42.877143 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.028045 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.125292 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.149637 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.186969 4690 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.196949 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.332037 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.468605 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.469134 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.632840 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.732932 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.733692 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.798149 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.801724 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.802218 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.871202 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.904989 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.933607 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.982136 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.996223 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 13:13:43 crc kubenswrapper[4690]: I1203 13:13:43.996631 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.056353 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.080032 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.102325 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.111838 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.231023 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.307683 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.328815 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.472381 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.482332 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.488786 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.537128 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.602108 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.647522 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.708531 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.724783 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.749734 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.822956 4690 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.868993 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.936430 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.941303 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 13:13:44 crc kubenswrapper[4690]: I1203 13:13:44.964382 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.017789 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.098547 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.127071 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.136983 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.207418 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.226903 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.291475 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.317937 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.389726 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.467504 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.524487 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.570572 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.588661 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.632795 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.668385 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.798564 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.882972 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.918144 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 13:13:45 crc kubenswrapper[4690]: I1203 13:13:45.935914 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.003191 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.030235 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.143533 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.171961 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.185494 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.297066 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.352368 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.478524 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.544913 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.601965 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.602122 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.728596 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.745991 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.755123 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.757094 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.895775 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.915750 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 13:13:46 crc kubenswrapper[4690]: I1203 13:13:46.974086 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.239250 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.547920 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.566440 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.605739 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.671389 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.726347 4690 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.833147 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.848425 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.888206 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.912597 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.986082 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 13:13:47 crc kubenswrapper[4690]: I1203 13:13:47.987721 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.008325 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.040483 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.173533 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.222955 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.311104 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.348123 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.378948 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.395078 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.413757 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.595502 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.661216 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.706770 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.733231 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.864116 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.917678 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.944394 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 13:13:48 crc kubenswrapper[4690]: I1203 13:13:48.994062 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.017547 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.072905 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.099819 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.175071 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.187737 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.225372 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.275177 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.280366 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.311035 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.313203 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.396006 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.409493 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.436215 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.498495 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.589577 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.607098 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.622229 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.675104 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.783639 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.810358 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.825640 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.861721 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.872177 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.877243 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.924677 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.931814 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.933211 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.950091 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 13:13:49 crc kubenswrapper[4690]: I1203 13:13:49.984061 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.072840 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.082577 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.088919 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.095967 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.135669 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.159969 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.177772 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.351651 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.370960 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.372433 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.455483 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.480018 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.514237 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.573410 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.762479 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.812345 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.845966 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.921931 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.939059 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.945986 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 13:13:50 crc kubenswrapper[4690]: I1203 13:13:50.989703 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.023820 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.024294 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.048099 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.157753 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.237593 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.277476 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.351429 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.560579 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.605371 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.609033 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.654582 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.661986 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.693002 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.705303 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.726969 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.754056 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.845837 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 13:13:51 crc kubenswrapper[4690]: I1203 13:13:51.965733 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.056483 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.123182 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.151528 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.242112 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.246739 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.247987 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.326341 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.568062 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.574044 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.580003 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.671816 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.696509 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.727403 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 13:13:52 crc kubenswrapper[4690]: I1203 13:13:52.908016 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.024620 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.148213 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.173998 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.197349 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.206018 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.279887 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.356633 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.378073 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.384285 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.605020 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.653294 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.714898 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.728977 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.756203 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.808113 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 13:13:53 crc kubenswrapper[4690]: I1203 13:13:53.907617 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.032026 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.112401 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.211410 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.304376 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.314825 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.355371 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.373543 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.409683 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.410312 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.506964 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.633753 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.851607 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 13:13:54 crc kubenswrapper[4690]: I1203 13:13:54.884434 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 13:13:55 crc kubenswrapper[4690]: I1203 13:13:55.117471 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 13:13:55 crc kubenswrapper[4690]: I1203 13:13:55.173278 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 13:13:55 crc kubenswrapper[4690]: I1203 13:13:55.193720 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 13:13:55 crc kubenswrapper[4690]: I1203 13:13:55.240353 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 13:13:55 crc kubenswrapper[4690]: I1203 13:13:55.377556 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 13:13:55 crc kubenswrapper[4690]: I1203 13:13:55.420253 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 13:13:55 crc kubenswrapper[4690]: I1203 13:13:55.830946 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 13:13:55 crc kubenswrapper[4690]: I1203 13:13:55.963356 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 13:13:56 crc kubenswrapper[4690]: I1203 13:13:56.049927 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 13:13:56 crc kubenswrapper[4690]: I1203 13:13:56.256426 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 13:13:56 crc kubenswrapper[4690]: I1203 13:13:56.474009 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 13:13:56 crc kubenswrapper[4690]: I1203 13:13:56.918253 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 13:13:57 crc kubenswrapper[4690]: I1203 13:13:57.920999 4690 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 13:13:58 crc kubenswrapper[4690]: I1203 13:13:58.810394 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.856322 4690 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.856749 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-drrqc" podStartSLOduration=60.23086325 podStartE2EDuration="2m31.856725953s" podCreationTimestamp="2025-12-03 13:11:40 +0000 UTC" firstStartedPulling="2025-12-03 13:11:43.420828674 +0000 UTC m=+149.401749117" lastFinishedPulling="2025-12-03 13:13:15.046691387 +0000 UTC m=+241.027611820" observedRunningTime="2025-12-03 13:13:30.072973202 +0000 UTC m=+256.053893635" watchObservedRunningTime="2025-12-03 13:14:11.856725953 +0000 UTC m=+297.837646386" Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.859421 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=64.859402117 podStartE2EDuration="1m4.859402117s" podCreationTimestamp="2025-12-03 13:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:13:30.092807319 +0000 UTC m=+256.073727752" watchObservedRunningTime="2025-12-03 13:14:11.859402117 +0000 UTC m=+297.840322550" Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.863960 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.864043 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t8hvp","openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:14:11 crc kubenswrapper[4690]: E1203 13:14:11.864466 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" containerName="installer" Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.864489 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" containerName="installer" Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.864716 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="455990f2-8fc3-4aa8-9a20-1c1f4d747d51" containerName="installer" Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.865327 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gknq8","openshift-marketplace/redhat-operators-drrqc","openshift-marketplace/redhat-marketplace-rvd94","openshift-marketplace/certified-operators-f8wnd","openshift-marketplace/redhat-operators-5j9q8","openshift-marketplace/marketplace-operator-79b997595-bhp9h"] Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.865636 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" podUID="4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" containerName="marketplace-operator" containerID="cri-o://07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5" gracePeriod=30 Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.865796 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f8wnd" podUID="48f93164-1544-42e2-82e0-60a681752349" containerName="registry-server" containerID="cri-o://5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f" gracePeriod=30 Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.866017 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.866017 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gknq8" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerName="registry-server" containerID="cri-o://95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0" gracePeriod=30 Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.866162 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rvd94" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerName="registry-server" containerID="cri-o://3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62" gracePeriod=30 Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.866401 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5j9q8" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="registry-server" containerID="cri-o://dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6" gracePeriod=30 Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.866592 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-drrqc" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" containerName="registry-server" containerID="cri-o://ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68" gracePeriod=30 Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.873107 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:14:11 crc kubenswrapper[4690]: I1203 13:14:11.926071 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=41.926049112 podStartE2EDuration="41.926049112s" podCreationTimestamp="2025-12-03 13:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:14:11.888363664 +0000 UTC m=+297.869284117" watchObservedRunningTime="2025-12-03 13:14:11.926049112 +0000 UTC m=+297.906969545" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.067605 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlfn9\" (UniqueName: \"kubernetes.io/projected/2b3e95ae-d895-45bf-9abd-73cf60e5f139-kube-api-access-zlfn9\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.068077 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2b3e95ae-d895-45bf-9abd-73cf60e5f139-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.068150 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b3e95ae-d895-45bf-9abd-73cf60e5f139-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.169586 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlfn9\" (UniqueName: \"kubernetes.io/projected/2b3e95ae-d895-45bf-9abd-73cf60e5f139-kube-api-access-zlfn9\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.169629 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2b3e95ae-d895-45bf-9abd-73cf60e5f139-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.169683 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b3e95ae-d895-45bf-9abd-73cf60e5f139-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.171417 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b3e95ae-d895-45bf-9abd-73cf60e5f139-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.183906 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2b3e95ae-d895-45bf-9abd-73cf60e5f139-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.192277 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlfn9\" (UniqueName: \"kubernetes.io/projected/2b3e95ae-d895-45bf-9abd-73cf60e5f139-kube-api-access-zlfn9\") pod \"marketplace-operator-79b997595-t8hvp\" (UID: \"2b3e95ae-d895-45bf-9abd-73cf60e5f139\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.210757 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.304256 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.349904 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.352623 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.369006 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373421 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxr4r\" (UniqueName: \"kubernetes.io/projected/48f93164-1544-42e2-82e0-60a681752349-kube-api-access-vxr4r\") pod \"48f93164-1544-42e2-82e0-60a681752349\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373487 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-catalog-content\") pod \"48f93164-1544-42e2-82e0-60a681752349\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373509 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g77k\" (UniqueName: \"kubernetes.io/projected/6a9984da-231b-45bf-8583-6394924d2e3e-kube-api-access-5g77k\") pod \"6a9984da-231b-45bf-8583-6394924d2e3e\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373533 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-689m7\" (UniqueName: \"kubernetes.io/projected/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-kube-api-access-689m7\") pod \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373557 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-catalog-content\") pod \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373590 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-utilities\") pod \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373608 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-utilities\") pod \"6a9984da-231b-45bf-8583-6394924d2e3e\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373636 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-operator-metrics\") pod \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373661 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-catalog-content\") pod \"6a9984da-231b-45bf-8583-6394924d2e3e\" (UID: \"6a9984da-231b-45bf-8583-6394924d2e3e\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373686 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9x9h\" (UniqueName: \"kubernetes.io/projected/c8c81207-0aa4-4489-a0e5-ed680f9436d2-kube-api-access-j9x9h\") pod \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\" (UID: \"c8c81207-0aa4-4489-a0e5-ed680f9436d2\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373702 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-utilities\") pod \"48f93164-1544-42e2-82e0-60a681752349\" (UID: \"48f93164-1544-42e2-82e0-60a681752349\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.373745 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-trusted-ca\") pod \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\" (UID: \"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.375908 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-utilities" (OuterVolumeSpecName: "utilities") pod "c8c81207-0aa4-4489-a0e5-ed680f9436d2" (UID: "c8c81207-0aa4-4489-a0e5-ed680f9436d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.377957 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" (UID: "4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.380043 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-utilities" (OuterVolumeSpecName: "utilities") pod "6a9984da-231b-45bf-8583-6394924d2e3e" (UID: "6a9984da-231b-45bf-8583-6394924d2e3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.382051 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48f93164-1544-42e2-82e0-60a681752349-kube-api-access-vxr4r" (OuterVolumeSpecName: "kube-api-access-vxr4r") pod "48f93164-1544-42e2-82e0-60a681752349" (UID: "48f93164-1544-42e2-82e0-60a681752349"). InnerVolumeSpecName "kube-api-access-vxr4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.382528 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-kube-api-access-689m7" (OuterVolumeSpecName: "kube-api-access-689m7") pod "4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" (UID: "4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9"). InnerVolumeSpecName "kube-api-access-689m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.390241 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a9984da-231b-45bf-8583-6394924d2e3e-kube-api-access-5g77k" (OuterVolumeSpecName: "kube-api-access-5g77k") pod "6a9984da-231b-45bf-8583-6394924d2e3e" (UID: "6a9984da-231b-45bf-8583-6394924d2e3e"). InnerVolumeSpecName "kube-api-access-5g77k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.394094 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.395256 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.397570 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-utilities" (OuterVolumeSpecName: "utilities") pod "48f93164-1544-42e2-82e0-60a681752349" (UID: "48f93164-1544-42e2-82e0-60a681752349"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.397589 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" (UID: "4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.403263 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c81207-0aa4-4489-a0e5-ed680f9436d2-kube-api-access-j9x9h" (OuterVolumeSpecName: "kube-api-access-j9x9h") pod "c8c81207-0aa4-4489-a0e5-ed680f9436d2" (UID: "c8c81207-0aa4-4489-a0e5-ed680f9436d2"). InnerVolumeSpecName "kube-api-access-j9x9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.471313 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8c81207-0aa4-4489-a0e5-ed680f9436d2" (UID: "c8c81207-0aa4-4489-a0e5-ed680f9436d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.474803 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-utilities\") pod \"c06438b0-cd29-4387-96fd-257f01c8fd11\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475018 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-catalog-content\") pod \"c06438b0-cd29-4387-96fd-257f01c8fd11\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475050 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-utilities\") pod \"9227b45f-65e8-4ed0-877a-fa41f765078f\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475113 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dnj7\" (UniqueName: \"kubernetes.io/projected/c06438b0-cd29-4387-96fd-257f01c8fd11-kube-api-access-6dnj7\") pod \"c06438b0-cd29-4387-96fd-257f01c8fd11\" (UID: \"c06438b0-cd29-4387-96fd-257f01c8fd11\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475147 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-catalog-content\") pod \"9227b45f-65e8-4ed0-877a-fa41f765078f\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475239 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2w69\" (UniqueName: \"kubernetes.io/projected/9227b45f-65e8-4ed0-877a-fa41f765078f-kube-api-access-h2w69\") pod \"9227b45f-65e8-4ed0-877a-fa41f765078f\" (UID: \"9227b45f-65e8-4ed0-877a-fa41f765078f\") " Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475472 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g77k\" (UniqueName: \"kubernetes.io/projected/6a9984da-231b-45bf-8583-6394924d2e3e-kube-api-access-5g77k\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475483 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-689m7\" (UniqueName: \"kubernetes.io/projected/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-kube-api-access-689m7\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475492 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475501 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81207-0aa4-4489-a0e5-ed680f9436d2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475510 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475522 4690 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475532 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9x9h\" (UniqueName: \"kubernetes.io/projected/c8c81207-0aa4-4489-a0e5-ed680f9436d2-kube-api-access-j9x9h\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475565 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475575 4690 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.475583 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxr4r\" (UniqueName: \"kubernetes.io/projected/48f93164-1544-42e2-82e0-60a681752349-kube-api-access-vxr4r\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.478205 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-utilities" (OuterVolumeSpecName: "utilities") pod "9227b45f-65e8-4ed0-877a-fa41f765078f" (UID: "9227b45f-65e8-4ed0-877a-fa41f765078f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.478797 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9227b45f-65e8-4ed0-877a-fa41f765078f-kube-api-access-h2w69" (OuterVolumeSpecName: "kube-api-access-h2w69") pod "9227b45f-65e8-4ed0-877a-fa41f765078f" (UID: "9227b45f-65e8-4ed0-877a-fa41f765078f"). InnerVolumeSpecName "kube-api-access-h2w69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.480824 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c06438b0-cd29-4387-96fd-257f01c8fd11-kube-api-access-6dnj7" (OuterVolumeSpecName: "kube-api-access-6dnj7") pod "c06438b0-cd29-4387-96fd-257f01c8fd11" (UID: "c06438b0-cd29-4387-96fd-257f01c8fd11"). InnerVolumeSpecName "kube-api-access-6dnj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.481474 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-utilities" (OuterVolumeSpecName: "utilities") pod "c06438b0-cd29-4387-96fd-257f01c8fd11" (UID: "c06438b0-cd29-4387-96fd-257f01c8fd11"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.492485 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48f93164-1544-42e2-82e0-60a681752349" (UID: "48f93164-1544-42e2-82e0-60a681752349"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.496546 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t8hvp"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.507077 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c06438b0-cd29-4387-96fd-257f01c8fd11" (UID: "c06438b0-cd29-4387-96fd-257f01c8fd11"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.536773 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a9984da-231b-45bf-8583-6394924d2e3e" (UID: "6a9984da-231b-45bf-8583-6394924d2e3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.577096 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9984da-231b-45bf-8583-6394924d2e3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.577143 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2w69\" (UniqueName: \"kubernetes.io/projected/9227b45f-65e8-4ed0-877a-fa41f765078f-kube-api-access-h2w69\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.577158 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.577170 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c06438b0-cd29-4387-96fd-257f01c8fd11-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.577182 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.577193 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f93164-1544-42e2-82e0-60a681752349-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.577204 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dnj7\" (UniqueName: \"kubernetes.io/projected/c06438b0-cd29-4387-96fd-257f01c8fd11-kube-api-access-6dnj7\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.603769 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9227b45f-65e8-4ed0-877a-fa41f765078f" (UID: "9227b45f-65e8-4ed0-877a-fa41f765078f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.678267 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9227b45f-65e8-4ed0-877a-fa41f765078f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.784142 4690 generic.go:334] "Generic (PLEG): container finished" podID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerID="3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62" exitCode=0 Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.784212 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvd94" event={"ID":"c06438b0-cd29-4387-96fd-257f01c8fd11","Type":"ContainerDied","Data":"3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.784255 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvd94" event={"ID":"c06438b0-cd29-4387-96fd-257f01c8fd11","Type":"ContainerDied","Data":"710b57da024e3184f624f9c18088c667c5bb17a794d4dff44edad1b4f5f43eb7"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.784276 4690 scope.go:117] "RemoveContainer" containerID="3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.784267 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvd94" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.786969 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerID="95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0" exitCode=0 Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.787021 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gknq8" event={"ID":"c8c81207-0aa4-4489-a0e5-ed680f9436d2","Type":"ContainerDied","Data":"95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.787048 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gknq8" event={"ID":"c8c81207-0aa4-4489-a0e5-ed680f9436d2","Type":"ContainerDied","Data":"ec3c9fa9755df3759defde07001b3e8b9b31368610515149e5893845921de9ac"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.787127 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gknq8" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.798474 4690 scope.go:117] "RemoveContainer" containerID="1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.799825 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" event={"ID":"2b3e95ae-d895-45bf-9abd-73cf60e5f139","Type":"ContainerStarted","Data":"7e7025c5b17a2e917d14501fe1c82a456dda144502c63b72d01a68ea4c7eb70b"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.799895 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" event={"ID":"2b3e95ae-d895-45bf-9abd-73cf60e5f139","Type":"ContainerStarted","Data":"b9310d4c4cd1f6411b657dbb133eea7124c4aa1d1ccec388a7192897575e248b"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.800564 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.802080 4690 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-t8hvp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.802136 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" podUID="2b3e95ae-d895-45bf-9abd-73cf60e5f139" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.804222 4690 generic.go:334] "Generic (PLEG): container finished" podID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerID="dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6" exitCode=0 Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.804273 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j9q8" event={"ID":"9227b45f-65e8-4ed0-877a-fa41f765078f","Type":"ContainerDied","Data":"dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.804341 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j9q8" event={"ID":"9227b45f-65e8-4ed0-877a-fa41f765078f","Type":"ContainerDied","Data":"68ec7708b1166b5a321fcf7b9c1d71c4f4157d3e32d5b10681108b81d8765439"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.804310 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5j9q8" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.810329 4690 generic.go:334] "Generic (PLEG): container finished" podID="4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" containerID="07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5" exitCode=0 Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.810579 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.810624 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" event={"ID":"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9","Type":"ContainerDied","Data":"07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.810679 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bhp9h" event={"ID":"4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9","Type":"ContainerDied","Data":"3f7cf4adeb60487ed4c9a8de6fef36edce0598d7dbf166c07142c11e85be70c6"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.816393 4690 generic.go:334] "Generic (PLEG): container finished" podID="48f93164-1544-42e2-82e0-60a681752349" containerID="5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f" exitCode=0 Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.816501 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8wnd" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.816508 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8wnd" event={"ID":"48f93164-1544-42e2-82e0-60a681752349","Type":"ContainerDied","Data":"5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.816659 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8wnd" event={"ID":"48f93164-1544-42e2-82e0-60a681752349","Type":"ContainerDied","Data":"0207cbbe10e776c52ffd3de95108219d95d3f7b3af03019fc7d8d01a063c73e8"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.819546 4690 generic.go:334] "Generic (PLEG): container finished" podID="6a9984da-231b-45bf-8583-6394924d2e3e" containerID="ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68" exitCode=0 Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.819942 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drrqc" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.820363 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drrqc" event={"ID":"6a9984da-231b-45bf-8583-6394924d2e3e","Type":"ContainerDied","Data":"ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.820433 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drrqc" event={"ID":"6a9984da-231b-45bf-8583-6394924d2e3e","Type":"ContainerDied","Data":"108bdf16dcc09a9b654d56b8fca71f24d01a4b13b989a4f146da8c4cd679e87d"} Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.822588 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" podStartSLOduration=6.822565572 podStartE2EDuration="6.822565572s" podCreationTimestamp="2025-12-03 13:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:14:12.81872852 +0000 UTC m=+298.799648963" watchObservedRunningTime="2025-12-03 13:14:12.822565572 +0000 UTC m=+298.803486005" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.833938 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gknq8"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.834278 4690 scope.go:117] "RemoveContainer" containerID="7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.836959 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gknq8"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.852564 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvd94"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.860710 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvd94"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.874935 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f8wnd"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.877711 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f8wnd"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.891429 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5j9q8"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.893994 4690 scope.go:117] "RemoveContainer" containerID="3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62" Dec 03 13:14:12 crc kubenswrapper[4690]: E1203 13:14:12.895801 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62\": container with ID starting with 3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62 not found: ID does not exist" containerID="3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.895851 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62"} err="failed to get container status \"3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62\": rpc error: code = NotFound desc = could not find container \"3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62\": container with ID starting with 3ac6bc47594956acead67685b3167617628987b974ca90b99632df50c5609a62 not found: ID does not exist" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.895899 4690 scope.go:117] "RemoveContainer" containerID="1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a" Dec 03 13:14:12 crc kubenswrapper[4690]: E1203 13:14:12.896724 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a\": container with ID starting with 1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a not found: ID does not exist" containerID="1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.896748 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a"} err="failed to get container status \"1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a\": rpc error: code = NotFound desc = could not find container \"1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a\": container with ID starting with 1feb7c5abc588753c8b70bfd6018057b367b3cd735b86cf71fd9ea943693ad2a not found: ID does not exist" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.896771 4690 scope.go:117] "RemoveContainer" containerID="7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e" Dec 03 13:14:12 crc kubenswrapper[4690]: E1203 13:14:12.897436 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e\": container with ID starting with 7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e not found: ID does not exist" containerID="7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.897459 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e"} err="failed to get container status \"7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e\": rpc error: code = NotFound desc = could not find container \"7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e\": container with ID starting with 7dd06666adaf8b7d2ffbf60eba676a5d21c3a724c606d0ed8d5c4fe7084fc56e not found: ID does not exist" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.897473 4690 scope.go:117] "RemoveContainer" containerID="95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.900618 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5j9q8"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.907727 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bhp9h"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.912068 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bhp9h"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.915977 4690 scope.go:117] "RemoveContainer" containerID="6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.916281 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-drrqc"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.919986 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-drrqc"] Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.931228 4690 scope.go:117] "RemoveContainer" containerID="bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.947101 4690 scope.go:117] "RemoveContainer" containerID="95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0" Dec 03 13:14:12 crc kubenswrapper[4690]: E1203 13:14:12.950456 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0\": container with ID starting with 95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0 not found: ID does not exist" containerID="95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.950514 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0"} err="failed to get container status \"95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0\": rpc error: code = NotFound desc = could not find container \"95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0\": container with ID starting with 95f0cf7e3564bdb977ad97a03908412ed705b60010635ff513935881b08acdb0 not found: ID does not exist" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.950559 4690 scope.go:117] "RemoveContainer" containerID="6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e" Dec 03 13:14:12 crc kubenswrapper[4690]: E1203 13:14:12.951069 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e\": container with ID starting with 6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e not found: ID does not exist" containerID="6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.951101 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e"} err="failed to get container status \"6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e\": rpc error: code = NotFound desc = could not find container \"6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e\": container with ID starting with 6de8d778e5d20b75643ed5f0048249d7e385b7f4aa4b966e881997c3531e9d7e not found: ID does not exist" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.951120 4690 scope.go:117] "RemoveContainer" containerID="bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508" Dec 03 13:14:12 crc kubenswrapper[4690]: E1203 13:14:12.951353 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508\": container with ID starting with bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508 not found: ID does not exist" containerID="bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.951381 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508"} err="failed to get container status \"bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508\": rpc error: code = NotFound desc = could not find container \"bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508\": container with ID starting with bfc44e176e3c51f2a7ba4a811787e4ab64c860cd92feb02874bf0ea8ffd61508 not found: ID does not exist" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.951399 4690 scope.go:117] "RemoveContainer" containerID="dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.965012 4690 scope.go:117] "RemoveContainer" containerID="f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7" Dec 03 13:14:12 crc kubenswrapper[4690]: I1203 13:14:12.985190 4690 scope.go:117] "RemoveContainer" containerID="ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.001152 4690 scope.go:117] "RemoveContainer" containerID="dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.001567 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6\": container with ID starting with dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6 not found: ID does not exist" containerID="dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.001616 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6"} err="failed to get container status \"dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6\": rpc error: code = NotFound desc = could not find container \"dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6\": container with ID starting with dd126417d02d9983fa60a7bf88bb109f1f15ad1b1e101ab256ba50eb3a76a6c6 not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.001652 4690 scope.go:117] "RemoveContainer" containerID="f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.002019 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7\": container with ID starting with f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7 not found: ID does not exist" containerID="f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.002048 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7"} err="failed to get container status \"f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7\": rpc error: code = NotFound desc = could not find container \"f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7\": container with ID starting with f9687212669c62d84ac6951434501d05744b523890b7f9a6f38aac36570fafa7 not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.002070 4690 scope.go:117] "RemoveContainer" containerID="ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.002315 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636\": container with ID starting with ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636 not found: ID does not exist" containerID="ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.002340 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636"} err="failed to get container status \"ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636\": rpc error: code = NotFound desc = could not find container \"ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636\": container with ID starting with ffe0770f3b6dc32d4f8377ccb2b10106e438c2afdc288e8cfe4a1fc73892b636 not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.002363 4690 scope.go:117] "RemoveContainer" containerID="07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.065987 4690 scope.go:117] "RemoveContainer" containerID="07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.066694 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5\": container with ID starting with 07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5 not found: ID does not exist" containerID="07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.066754 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5"} err="failed to get container status \"07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5\": rpc error: code = NotFound desc = could not find container \"07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5\": container with ID starting with 07f5ed0fdeedc054923f99eabf671128228f6522570f115b41e7c9b0a357d7e5 not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.066789 4690 scope.go:117] "RemoveContainer" containerID="5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.080733 4690 scope.go:117] "RemoveContainer" containerID="64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.097608 4690 scope.go:117] "RemoveContainer" containerID="9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.114265 4690 scope.go:117] "RemoveContainer" containerID="5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.114822 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f\": container with ID starting with 5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f not found: ID does not exist" containerID="5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.114895 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f"} err="failed to get container status \"5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f\": rpc error: code = NotFound desc = could not find container \"5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f\": container with ID starting with 5ee9c2fafa54ecf3b8d9097e0b6e5acef94f7a791ee3caa6cc98dc6ddcadd86f not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.114934 4690 scope.go:117] "RemoveContainer" containerID="64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.115338 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf\": container with ID starting with 64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf not found: ID does not exist" containerID="64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.115398 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf"} err="failed to get container status \"64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf\": rpc error: code = NotFound desc = could not find container \"64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf\": container with ID starting with 64f5d428cfca8ff0b77c5ff1bc8fe8741f5eee82bfad8207def4ed67eefcb7cf not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.115419 4690 scope.go:117] "RemoveContainer" containerID="9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.115822 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879\": container with ID starting with 9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879 not found: ID does not exist" containerID="9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.115878 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879"} err="failed to get container status \"9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879\": rpc error: code = NotFound desc = could not find container \"9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879\": container with ID starting with 9e13065e27537ff4f0a057fbf3e42cbdadbc7f4fec2c0b7a895fdcc7dc413879 not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.115898 4690 scope.go:117] "RemoveContainer" containerID="ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.127568 4690 scope.go:117] "RemoveContainer" containerID="7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.141362 4690 scope.go:117] "RemoveContainer" containerID="7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.158017 4690 scope.go:117] "RemoveContainer" containerID="ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.158766 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68\": container with ID starting with ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68 not found: ID does not exist" containerID="ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.158805 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68"} err="failed to get container status \"ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68\": rpc error: code = NotFound desc = could not find container \"ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68\": container with ID starting with ffe218f2ddbfee31dac9c6a50bfd00af22b6318c812aca7d2e9e904e84b68f68 not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.158834 4690 scope.go:117] "RemoveContainer" containerID="7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.159412 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e\": container with ID starting with 7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e not found: ID does not exist" containerID="7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.159482 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e"} err="failed to get container status \"7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e\": rpc error: code = NotFound desc = could not find container \"7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e\": container with ID starting with 7f07d8cb797b63c6998d2643547c1b00f3580074914852d9041d5ca921358c1e not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.159521 4690 scope.go:117] "RemoveContainer" containerID="7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26" Dec 03 13:14:13 crc kubenswrapper[4690]: E1203 13:14:13.159970 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26\": container with ID starting with 7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26 not found: ID does not exist" containerID="7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.160004 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26"} err="failed to get container status \"7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26\": rpc error: code = NotFound desc = could not find container \"7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26\": container with ID starting with 7bc9016ae116ff08778ed9a7d0a719a51dbfa29173c2442bc1ddc648bcae7c26 not found: ID does not exist" Dec 03 13:14:13 crc kubenswrapper[4690]: I1203 13:14:13.843542 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-t8hvp" Dec 03 13:14:14 crc kubenswrapper[4690]: I1203 13:14:14.322236 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48f93164-1544-42e2-82e0-60a681752349" path="/var/lib/kubelet/pods/48f93164-1544-42e2-82e0-60a681752349/volumes" Dec 03 13:14:14 crc kubenswrapper[4690]: I1203 13:14:14.323133 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" path="/var/lib/kubelet/pods/4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9/volumes" Dec 03 13:14:14 crc kubenswrapper[4690]: I1203 13:14:14.323621 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" path="/var/lib/kubelet/pods/6a9984da-231b-45bf-8583-6394924d2e3e/volumes" Dec 03 13:14:14 crc kubenswrapper[4690]: I1203 13:14:14.324639 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" path="/var/lib/kubelet/pods/9227b45f-65e8-4ed0-877a-fa41f765078f/volumes" Dec 03 13:14:14 crc kubenswrapper[4690]: I1203 13:14:14.325250 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" path="/var/lib/kubelet/pods/c06438b0-cd29-4387-96fd-257f01c8fd11/volumes" Dec 03 13:14:14 crc kubenswrapper[4690]: I1203 13:14:14.326216 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" path="/var/lib/kubelet/pods/c8c81207-0aa4-4489-a0e5-ed680f9436d2/volumes" Dec 03 13:14:15 crc kubenswrapper[4690]: I1203 13:14:15.381042 4690 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 13:14:15 crc kubenswrapper[4690]: I1203 13:14:15.381303 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://bd2b6627fd9da4c9bcaddf8ea22923cbefe6a05bbd1559eb000aeebeae998a6f" gracePeriod=5 Dec 03 13:14:20 crc kubenswrapper[4690]: I1203 13:14:20.877525 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 13:14:20 crc kubenswrapper[4690]: I1203 13:14:20.879245 4690 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="bd2b6627fd9da4c9bcaddf8ea22923cbefe6a05bbd1559eb000aeebeae998a6f" exitCode=137 Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.006734 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.007191 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122240 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122329 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122388 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122429 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122448 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122434 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122467 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122499 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122527 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122894 4690 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122911 4690 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122921 4690 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.122929 4690 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.132044 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.223665 4690 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.886897 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.886989 4690 scope.go:117] "RemoveContainer" containerID="bd2b6627fd9da4c9bcaddf8ea22923cbefe6a05bbd1559eb000aeebeae998a6f" Dec 03 13:14:21 crc kubenswrapper[4690]: I1203 13:14:21.887058 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:14:22 crc kubenswrapper[4690]: I1203 13:14:22.321236 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 13:14:22 crc kubenswrapper[4690]: I1203 13:14:22.321721 4690 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 03 13:14:22 crc kubenswrapper[4690]: I1203 13:14:22.331392 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 13:14:22 crc kubenswrapper[4690]: I1203 13:14:22.331481 4690 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="7f23a7ea-3743-42bf-ac61-218d71c64bd9" Dec 03 13:14:22 crc kubenswrapper[4690]: I1203 13:14:22.334759 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 13:14:22 crc kubenswrapper[4690]: I1203 13:14:22.334809 4690 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="7f23a7ea-3743-42bf-ac61-218d71c64bd9" Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.367569 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jdlts"] Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.368658 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" podUID="4975e4cf-5be0-40e1-a4b0-2db6ec43c207" containerName="controller-manager" containerID="cri-o://02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b" gracePeriod=30 Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.465025 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788"] Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.465650 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" podUID="da998e2c-9c36-4f73-a839-d81ad59c955d" containerName="route-controller-manager" containerID="cri-o://32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984" gracePeriod=30 Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.777765 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.852987 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.929683 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca\") pod \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.929752 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles\") pod \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.929822 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrjp8\" (UniqueName: \"kubernetes.io/projected/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-kube-api-access-hrjp8\") pod \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.929859 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert\") pod \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.929958 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config\") pod \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\" (UID: \"4975e4cf-5be0-40e1-a4b0-2db6ec43c207\") " Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.930687 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca" (OuterVolumeSpecName: "client-ca") pod "4975e4cf-5be0-40e1-a4b0-2db6ec43c207" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.931294 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4975e4cf-5be0-40e1-a4b0-2db6ec43c207" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.931312 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config" (OuterVolumeSpecName: "config") pod "4975e4cf-5be0-40e1-a4b0-2db6ec43c207" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.936085 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4975e4cf-5be0-40e1-a4b0-2db6ec43c207" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:14:52 crc kubenswrapper[4690]: I1203 13:14:52.936379 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-kube-api-access-hrjp8" (OuterVolumeSpecName: "kube-api-access-hrjp8") pod "4975e4cf-5be0-40e1-a4b0-2db6ec43c207" (UID: "4975e4cf-5be0-40e1-a4b0-2db6ec43c207"). InnerVolumeSpecName "kube-api-access-hrjp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.031132 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-client-ca\") pod \"da998e2c-9c36-4f73-a839-d81ad59c955d\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.031241 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da998e2c-9c36-4f73-a839-d81ad59c955d-serving-cert\") pod \"da998e2c-9c36-4f73-a839-d81ad59c955d\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.031658 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-config\") pod \"da998e2c-9c36-4f73-a839-d81ad59c955d\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.031750 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdbx9\" (UniqueName: \"kubernetes.io/projected/da998e2c-9c36-4f73-a839-d81ad59c955d-kube-api-access-mdbx9\") pod \"da998e2c-9c36-4f73-a839-d81ad59c955d\" (UID: \"da998e2c-9c36-4f73-a839-d81ad59c955d\") " Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.032177 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-client-ca" (OuterVolumeSpecName: "client-ca") pod "da998e2c-9c36-4f73-a839-d81ad59c955d" (UID: "da998e2c-9c36-4f73-a839-d81ad59c955d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.032253 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.032292 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.032319 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrjp8\" (UniqueName: \"kubernetes.io/projected/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-kube-api-access-hrjp8\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.032344 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.032366 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4975e4cf-5be0-40e1-a4b0-2db6ec43c207-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.032653 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-config" (OuterVolumeSpecName: "config") pod "da998e2c-9c36-4f73-a839-d81ad59c955d" (UID: "da998e2c-9c36-4f73-a839-d81ad59c955d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.035708 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da998e2c-9c36-4f73-a839-d81ad59c955d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "da998e2c-9c36-4f73-a839-d81ad59c955d" (UID: "da998e2c-9c36-4f73-a839-d81ad59c955d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.035782 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da998e2c-9c36-4f73-a839-d81ad59c955d-kube-api-access-mdbx9" (OuterVolumeSpecName: "kube-api-access-mdbx9") pod "da998e2c-9c36-4f73-a839-d81ad59c955d" (UID: "da998e2c-9c36-4f73-a839-d81ad59c955d"). InnerVolumeSpecName "kube-api-access-mdbx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.044782 4690 generic.go:334] "Generic (PLEG): container finished" podID="4975e4cf-5be0-40e1-a4b0-2db6ec43c207" containerID="02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b" exitCode=0 Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.044891 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" event={"ID":"4975e4cf-5be0-40e1-a4b0-2db6ec43c207","Type":"ContainerDied","Data":"02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b"} Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.044958 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" event={"ID":"4975e4cf-5be0-40e1-a4b0-2db6ec43c207","Type":"ContainerDied","Data":"8993a0d43ac5a838bc415a07ffc63bdfe6b9829c222f3fb569565093798c5e2d"} Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.044982 4690 scope.go:117] "RemoveContainer" containerID="02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.045158 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jdlts" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.048526 4690 generic.go:334] "Generic (PLEG): container finished" podID="da998e2c-9c36-4f73-a839-d81ad59c955d" containerID="32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984" exitCode=0 Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.048572 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.048582 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" event={"ID":"da998e2c-9c36-4f73-a839-d81ad59c955d","Type":"ContainerDied","Data":"32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984"} Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.048636 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788" event={"ID":"da998e2c-9c36-4f73-a839-d81ad59c955d","Type":"ContainerDied","Data":"75664de918e11a42228c2f740e69b9a6134cc7db5210d86a467a0d08cb80b67e"} Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.062531 4690 scope.go:117] "RemoveContainer" containerID="02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.062982 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b\": container with ID starting with 02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b not found: ID does not exist" containerID="02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.063636 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b"} err="failed to get container status \"02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b\": rpc error: code = NotFound desc = could not find container \"02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b\": container with ID starting with 02c2c7fe9135204bf341e59773c96dcd2a2e0a628e3319b13810b82874daec6b not found: ID does not exist" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.063744 4690 scope.go:117] "RemoveContainer" containerID="32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.081113 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jdlts"] Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.083840 4690 scope.go:117] "RemoveContainer" containerID="32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.085012 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jdlts"] Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.085534 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984\": container with ID starting with 32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984 not found: ID does not exist" containerID="32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.085574 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984"} err="failed to get container status \"32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984\": rpc error: code = NotFound desc = could not find container \"32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984\": container with ID starting with 32abc8bcc8ba0a87d6fd5d46f0c0dd6289925d13d5f3294f1eaff2d043cd5984 not found: ID does not exist" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.112966 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788"] Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.116388 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d5788"] Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.133583 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.133625 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdbx9\" (UniqueName: \"kubernetes.io/projected/da998e2c-9c36-4f73-a839-d81ad59c955d-kube-api-access-mdbx9\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.133642 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da998e2c-9c36-4f73-a839-d81ad59c955d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.133679 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da998e2c-9c36-4f73-a839-d81ad59c955d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209416 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c"] Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209671 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209686 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209701 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209709 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209723 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209731 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209740 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" containerName="marketplace-operator" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209748 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" containerName="marketplace-operator" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209758 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209766 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209776 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209784 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209798 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da998e2c-9c36-4f73-a839-d81ad59c955d" containerName="route-controller-manager" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209806 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="da998e2c-9c36-4f73-a839-d81ad59c955d" containerName="route-controller-manager" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209816 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f93164-1544-42e2-82e0-60a681752349" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209824 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f93164-1544-42e2-82e0-60a681752349" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209832 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209840 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209851 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f93164-1544-42e2-82e0-60a681752349" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209859 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f93164-1544-42e2-82e0-60a681752349" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209907 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209915 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209924 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209932 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209943 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209950 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209963 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4975e4cf-5be0-40e1-a4b0-2db6ec43c207" containerName="controller-manager" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209970 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4975e4cf-5be0-40e1-a4b0-2db6ec43c207" containerName="controller-manager" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209980 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.209987 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerName="extract-utilities" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.209999 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210008 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.210020 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f93164-1544-42e2-82e0-60a681752349" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210027 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f93164-1544-42e2-82e0-60a681752349" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.210035 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210043 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerName="extract-content" Dec 03 13:14:53 crc kubenswrapper[4690]: E1203 13:14:53.210054 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210061 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210188 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c1b6b9f-46b2-49c9-a8f5-5821ab5387b9" containerName="marketplace-operator" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210202 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="48f93164-1544-42e2-82e0-60a681752349" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210214 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c06438b0-cd29-4387-96fd-257f01c8fd11" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210225 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="9227b45f-65e8-4ed0-877a-fa41f765078f" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210236 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c81207-0aa4-4489-a0e5-ed680f9436d2" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210247 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="4975e4cf-5be0-40e1-a4b0-2db6ec43c207" containerName="controller-manager" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210258 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a9984da-231b-45bf-8583-6394924d2e3e" containerName="registry-server" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210271 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210281 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="da998e2c-9c36-4f73-a839-d81ad59c955d" containerName="route-controller-manager" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.210970 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.213276 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.213492 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.213724 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.214383 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.214571 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.214647 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.222190 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c"] Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.224603 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.284886 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z"] Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.285581 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.287388 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.287512 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.288301 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.288431 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.288515 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.290709 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.293437 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z"] Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.336230 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-proxy-ca-bundles\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.336411 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69fg6\" (UniqueName: \"kubernetes.io/projected/276f20a9-9241-4ea1-a33f-7def78f05efd-kube-api-access-69fg6\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.336505 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-client-ca\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.336538 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276f20a9-9241-4ea1-a33f-7def78f05efd-serving-cert\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.336584 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-config\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438180 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-client-ca\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438234 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276f20a9-9241-4ea1-a33f-7def78f05efd-serving-cert\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438272 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-config\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438311 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-config\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438340 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-proxy-ca-bundles\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438363 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-client-ca\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438389 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwnlk\" (UniqueName: \"kubernetes.io/projected/ed128bdd-2a3f-4286-9ea8-295681e5cc89-kube-api-access-bwnlk\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438464 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69fg6\" (UniqueName: \"kubernetes.io/projected/276f20a9-9241-4ea1-a33f-7def78f05efd-kube-api-access-69fg6\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.438491 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed128bdd-2a3f-4286-9ea8-295681e5cc89-serving-cert\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.439313 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-client-ca\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.440057 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-config\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.440134 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-proxy-ca-bundles\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.442762 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276f20a9-9241-4ea1-a33f-7def78f05efd-serving-cert\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.455578 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69fg6\" (UniqueName: \"kubernetes.io/projected/276f20a9-9241-4ea1-a33f-7def78f05efd-kube-api-access-69fg6\") pod \"controller-manager-6cdffcbc8f-ltm7c\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.531252 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.539973 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed128bdd-2a3f-4286-9ea8-295681e5cc89-serving-cert\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.540022 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-config\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.540072 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-client-ca\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.540105 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwnlk\" (UniqueName: \"kubernetes.io/projected/ed128bdd-2a3f-4286-9ea8-295681e5cc89-kube-api-access-bwnlk\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.541406 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-client-ca\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.541933 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-config\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.544522 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed128bdd-2a3f-4286-9ea8-295681e5cc89-serving-cert\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.556748 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwnlk\" (UniqueName: \"kubernetes.io/projected/ed128bdd-2a3f-4286-9ea8-295681e5cc89-kube-api-access-bwnlk\") pod \"route-controller-manager-557bb68cfc-lpw6z\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.600527 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.723830 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c"] Dec 03 13:14:53 crc kubenswrapper[4690]: I1203 13:14:53.848276 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z"] Dec 03 13:14:53 crc kubenswrapper[4690]: W1203 13:14:53.852211 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded128bdd_2a3f_4286_9ea8_295681e5cc89.slice/crio-4fce32aef77fd75c5628bf2e21bbe065a45688d4856c6fb4186b180cce7e44d8 WatchSource:0}: Error finding container 4fce32aef77fd75c5628bf2e21bbe065a45688d4856c6fb4186b180cce7e44d8: Status 404 returned error can't find the container with id 4fce32aef77fd75c5628bf2e21bbe065a45688d4856c6fb4186b180cce7e44d8 Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.055110 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" event={"ID":"276f20a9-9241-4ea1-a33f-7def78f05efd","Type":"ContainerStarted","Data":"0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5"} Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.055550 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.055570 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" event={"ID":"276f20a9-9241-4ea1-a33f-7def78f05efd","Type":"ContainerStarted","Data":"87effcec771a21ffdbf33a58226d32ba4f237de8e0c3487427fd16238c9c5c11"} Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.056314 4690 patch_prober.go:28] interesting pod/controller-manager-6cdffcbc8f-ltm7c container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.056353 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" podUID="276f20a9-9241-4ea1-a33f-7def78f05efd" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.058135 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" event={"ID":"ed128bdd-2a3f-4286-9ea8-295681e5cc89","Type":"ContainerStarted","Data":"19c8b9caa4f383145f796dcfc5efe8c121342842a393d5bb2ac404cf915b0f40"} Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.058164 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" event={"ID":"ed128bdd-2a3f-4286-9ea8-295681e5cc89","Type":"ContainerStarted","Data":"4fce32aef77fd75c5628bf2e21bbe065a45688d4856c6fb4186b180cce7e44d8"} Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.058342 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.059684 4690 patch_prober.go:28] interesting pod/route-controller-manager-557bb68cfc-lpw6z container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.059724 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" podUID="ed128bdd-2a3f-4286-9ea8-295681e5cc89" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.097185 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" podStartSLOduration=1.097159982 podStartE2EDuration="1.097159982s" podCreationTimestamp="2025-12-03 13:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:14:54.079536454 +0000 UTC m=+340.060456887" watchObservedRunningTime="2025-12-03 13:14:54.097159982 +0000 UTC m=+340.078080425" Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.101621 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" podStartSLOduration=1.101607917 podStartE2EDuration="1.101607917s" podCreationTimestamp="2025-12-03 13:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:14:54.100513154 +0000 UTC m=+340.081433597" watchObservedRunningTime="2025-12-03 13:14:54.101607917 +0000 UTC m=+340.082528350" Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.322380 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4975e4cf-5be0-40e1-a4b0-2db6ec43c207" path="/var/lib/kubelet/pods/4975e4cf-5be0-40e1-a4b0-2db6ec43c207/volumes" Dec 03 13:14:54 crc kubenswrapper[4690]: I1203 13:14:54.323109 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da998e2c-9c36-4f73-a839-d81ad59c955d" path="/var/lib/kubelet/pods/da998e2c-9c36-4f73-a839-d81ad59c955d/volumes" Dec 03 13:14:55 crc kubenswrapper[4690]: I1203 13:14:55.066690 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:14:55 crc kubenswrapper[4690]: I1203 13:14:55.067403 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.164053 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj"] Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.166052 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.167107 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj"] Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.173195 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.173362 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.338403 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9235e7b-0d79-471b-96fa-08c36c7f421f-secret-volume\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.338545 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgfzj\" (UniqueName: \"kubernetes.io/projected/e9235e7b-0d79-471b-96fa-08c36c7f421f-kube-api-access-hgfzj\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.338629 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9235e7b-0d79-471b-96fa-08c36c7f421f-config-volume\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.440090 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgfzj\" (UniqueName: \"kubernetes.io/projected/e9235e7b-0d79-471b-96fa-08c36c7f421f-kube-api-access-hgfzj\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.440235 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9235e7b-0d79-471b-96fa-08c36c7f421f-config-volume\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.440283 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9235e7b-0d79-471b-96fa-08c36c7f421f-secret-volume\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.442551 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9235e7b-0d79-471b-96fa-08c36c7f421f-config-volume\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.447534 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9235e7b-0d79-471b-96fa-08c36c7f421f-secret-volume\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.460858 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgfzj\" (UniqueName: \"kubernetes.io/projected/e9235e7b-0d79-471b-96fa-08c36c7f421f-kube-api-access-hgfzj\") pod \"collect-profiles-29412795-64crj\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.487537 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:00 crc kubenswrapper[4690]: I1203 13:15:00.920140 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj"] Dec 03 13:15:01 crc kubenswrapper[4690]: I1203 13:15:01.098048 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" event={"ID":"e9235e7b-0d79-471b-96fa-08c36c7f421f","Type":"ContainerStarted","Data":"20a3d77817fc21a96e4867c8a1c2452347520db2c20454b888dafbb88e6e6e01"} Dec 03 13:15:01 crc kubenswrapper[4690]: I1203 13:15:01.098448 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" event={"ID":"e9235e7b-0d79-471b-96fa-08c36c7f421f","Type":"ContainerStarted","Data":"cb511d495e0c6af22129bd9fcbe9201e5d70bd9363ead366c92d9e0093962fb6"} Dec 03 13:15:01 crc kubenswrapper[4690]: I1203 13:15:01.113339 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" podStartSLOduration=1.113320521 podStartE2EDuration="1.113320521s" podCreationTimestamp="2025-12-03 13:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:15:01.111623925 +0000 UTC m=+347.092544378" watchObservedRunningTime="2025-12-03 13:15:01.113320521 +0000 UTC m=+347.094240954" Dec 03 13:15:02 crc kubenswrapper[4690]: I1203 13:15:02.104818 4690 generic.go:334] "Generic (PLEG): container finished" podID="e9235e7b-0d79-471b-96fa-08c36c7f421f" containerID="20a3d77817fc21a96e4867c8a1c2452347520db2c20454b888dafbb88e6e6e01" exitCode=0 Dec 03 13:15:02 crc kubenswrapper[4690]: I1203 13:15:02.104910 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" event={"ID":"e9235e7b-0d79-471b-96fa-08c36c7f421f","Type":"ContainerDied","Data":"20a3d77817fc21a96e4867c8a1c2452347520db2c20454b888dafbb88e6e6e01"} Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.439662 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.627631 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9235e7b-0d79-471b-96fa-08c36c7f421f-config-volume\") pod \"e9235e7b-0d79-471b-96fa-08c36c7f421f\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.627723 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9235e7b-0d79-471b-96fa-08c36c7f421f-secret-volume\") pod \"e9235e7b-0d79-471b-96fa-08c36c7f421f\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.627774 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgfzj\" (UniqueName: \"kubernetes.io/projected/e9235e7b-0d79-471b-96fa-08c36c7f421f-kube-api-access-hgfzj\") pod \"e9235e7b-0d79-471b-96fa-08c36c7f421f\" (UID: \"e9235e7b-0d79-471b-96fa-08c36c7f421f\") " Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.628469 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9235e7b-0d79-471b-96fa-08c36c7f421f-config-volume" (OuterVolumeSpecName: "config-volume") pod "e9235e7b-0d79-471b-96fa-08c36c7f421f" (UID: "e9235e7b-0d79-471b-96fa-08c36c7f421f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.634699 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9235e7b-0d79-471b-96fa-08c36c7f421f-kube-api-access-hgfzj" (OuterVolumeSpecName: "kube-api-access-hgfzj") pod "e9235e7b-0d79-471b-96fa-08c36c7f421f" (UID: "e9235e7b-0d79-471b-96fa-08c36c7f421f"). InnerVolumeSpecName "kube-api-access-hgfzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.638932 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9235e7b-0d79-471b-96fa-08c36c7f421f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e9235e7b-0d79-471b-96fa-08c36c7f421f" (UID: "e9235e7b-0d79-471b-96fa-08c36c7f421f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.729453 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9235e7b-0d79-471b-96fa-08c36c7f421f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.729526 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgfzj\" (UniqueName: \"kubernetes.io/projected/e9235e7b-0d79-471b-96fa-08c36c7f421f-kube-api-access-hgfzj\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:03 crc kubenswrapper[4690]: I1203 13:15:03.729539 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9235e7b-0d79-471b-96fa-08c36c7f421f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:04 crc kubenswrapper[4690]: I1203 13:15:04.116833 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" event={"ID":"e9235e7b-0d79-471b-96fa-08c36c7f421f","Type":"ContainerDied","Data":"cb511d495e0c6af22129bd9fcbe9201e5d70bd9363ead366c92d9e0093962fb6"} Dec 03 13:15:04 crc kubenswrapper[4690]: I1203 13:15:04.116900 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb511d495e0c6af22129bd9fcbe9201e5d70bd9363ead366c92d9e0093962fb6" Dec 03 13:15:04 crc kubenswrapper[4690]: I1203 13:15:04.116921 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.024254 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2l4wj"] Dec 03 13:15:10 crc kubenswrapper[4690]: E1203 13:15:10.024794 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9235e7b-0d79-471b-96fa-08c36c7f421f" containerName="collect-profiles" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.024807 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9235e7b-0d79-471b-96fa-08c36c7f421f" containerName="collect-profiles" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.024927 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9235e7b-0d79-471b-96fa-08c36c7f421f" containerName="collect-profiles" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.025386 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.037725 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2l4wj"] Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.222129 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxbbd\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-kube-api-access-wxbbd\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.222693 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.222714 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-trusted-ca\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.222754 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-registry-certificates\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.222778 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-registry-tls\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.222803 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.222963 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.222980 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-bound-sa-token\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.246097 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.323654 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-registry-tls\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.323753 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.323826 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-bound-sa-token\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.324268 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxbbd\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-kube-api-access-wxbbd\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.324502 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-trusted-ca\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.324532 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.325011 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.325909 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-trusted-ca\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.325971 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-registry-certificates\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.326936 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-registry-certificates\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.331554 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-registry-tls\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.332726 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.344141 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-bound-sa-token\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.345337 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxbbd\" (UniqueName: \"kubernetes.io/projected/00a6b5de-a44e-4df5-89b9-7b8c7916d82e-kube-api-access-wxbbd\") pod \"image-registry-66df7c8f76-2l4wj\" (UID: \"00a6b5de-a44e-4df5-89b9-7b8c7916d82e\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.346649 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:10 crc kubenswrapper[4690]: I1203 13:15:10.762158 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2l4wj"] Dec 03 13:15:11 crc kubenswrapper[4690]: I1203 13:15:11.155519 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" event={"ID":"00a6b5de-a44e-4df5-89b9-7b8c7916d82e","Type":"ContainerStarted","Data":"2c242edb90e37f92e49cb77c80a2f2cd5037ddf8b55cfc01df027917c890b43e"} Dec 03 13:15:11 crc kubenswrapper[4690]: I1203 13:15:11.155577 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" event={"ID":"00a6b5de-a44e-4df5-89b9-7b8c7916d82e","Type":"ContainerStarted","Data":"03d64d9045c73c6756fdcb23f53d31494290defd5e9e0eb9bf997cf493d598c8"} Dec 03 13:15:11 crc kubenswrapper[4690]: I1203 13:15:11.156636 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:12 crc kubenswrapper[4690]: I1203 13:15:12.754607 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" podStartSLOduration=2.754579027 podStartE2EDuration="2.754579027s" podCreationTimestamp="2025-12-03 13:15:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:15:11.181263433 +0000 UTC m=+357.162183866" watchObservedRunningTime="2025-12-03 13:15:12.754579027 +0000 UTC m=+358.735499460" Dec 03 13:15:12 crc kubenswrapper[4690]: I1203 13:15:12.755718 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z"] Dec 03 13:15:12 crc kubenswrapper[4690]: I1203 13:15:12.755936 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" podUID="ed128bdd-2a3f-4286-9ea8-295681e5cc89" containerName="route-controller-manager" containerID="cri-o://19c8b9caa4f383145f796dcfc5efe8c121342842a393d5bb2ac404cf915b0f40" gracePeriod=30 Dec 03 13:15:13 crc kubenswrapper[4690]: I1203 13:15:13.603225 4690 patch_prober.go:28] interesting pod/route-controller-manager-557bb68cfc-lpw6z container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Dec 03 13:15:13 crc kubenswrapper[4690]: I1203 13:15:13.603288 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" podUID="ed128bdd-2a3f-4286-9ea8-295681e5cc89" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.173654 4690 generic.go:334] "Generic (PLEG): container finished" podID="ed128bdd-2a3f-4286-9ea8-295681e5cc89" containerID="19c8b9caa4f383145f796dcfc5efe8c121342842a393d5bb2ac404cf915b0f40" exitCode=0 Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.174105 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" event={"ID":"ed128bdd-2a3f-4286-9ea8-295681e5cc89","Type":"ContainerDied","Data":"19c8b9caa4f383145f796dcfc5efe8c121342842a393d5bb2ac404cf915b0f40"} Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.363439 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.403394 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w"] Dec 03 13:15:14 crc kubenswrapper[4690]: E1203 13:15:14.403731 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed128bdd-2a3f-4286-9ea8-295681e5cc89" containerName="route-controller-manager" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.403749 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed128bdd-2a3f-4286-9ea8-295681e5cc89" containerName="route-controller-manager" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.403931 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed128bdd-2a3f-4286-9ea8-295681e5cc89" containerName="route-controller-manager" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.404360 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.413206 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w"] Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.480916 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed128bdd-2a3f-4286-9ea8-295681e5cc89-serving-cert\") pod \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.481006 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-client-ca\") pod \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.481134 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwnlk\" (UniqueName: \"kubernetes.io/projected/ed128bdd-2a3f-4286-9ea8-295681e5cc89-kube-api-access-bwnlk\") pod \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.481186 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-config\") pod \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\" (UID: \"ed128bdd-2a3f-4286-9ea8-295681e5cc89\") " Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.481606 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-client-ca" (OuterVolumeSpecName: "client-ca") pod "ed128bdd-2a3f-4286-9ea8-295681e5cc89" (UID: "ed128bdd-2a3f-4286-9ea8-295681e5cc89"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.482428 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-config" (OuterVolumeSpecName: "config") pod "ed128bdd-2a3f-4286-9ea8-295681e5cc89" (UID: "ed128bdd-2a3f-4286-9ea8-295681e5cc89"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.487283 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed128bdd-2a3f-4286-9ea8-295681e5cc89-kube-api-access-bwnlk" (OuterVolumeSpecName: "kube-api-access-bwnlk") pod "ed128bdd-2a3f-4286-9ea8-295681e5cc89" (UID: "ed128bdd-2a3f-4286-9ea8-295681e5cc89"). InnerVolumeSpecName "kube-api-access-bwnlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.487733 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed128bdd-2a3f-4286-9ea8-295681e5cc89-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ed128bdd-2a3f-4286-9ea8-295681e5cc89" (UID: "ed128bdd-2a3f-4286-9ea8-295681e5cc89"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.582534 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-client-ca\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.583004 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-config\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.583203 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69npj\" (UniqueName: \"kubernetes.io/projected/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-kube-api-access-69npj\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.583376 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-serving-cert\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.583515 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.583608 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed128bdd-2a3f-4286-9ea8-295681e5cc89-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.583684 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed128bdd-2a3f-4286-9ea8-295681e5cc89-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.583762 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwnlk\" (UniqueName: \"kubernetes.io/projected/ed128bdd-2a3f-4286-9ea8-295681e5cc89-kube-api-access-bwnlk\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.685421 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-config\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.685502 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69npj\" (UniqueName: \"kubernetes.io/projected/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-kube-api-access-69npj\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.685546 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-serving-cert\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.685578 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-client-ca\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.687164 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-config\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.687507 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-client-ca\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.694146 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-serving-cert\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.707665 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69npj\" (UniqueName: \"kubernetes.io/projected/cc080a74-f23e-4f89-9fc0-8cab2012a2f4-kube-api-access-69npj\") pod \"route-controller-manager-86cbd6df86-2rw5w\" (UID: \"cc080a74-f23e-4f89-9fc0-8cab2012a2f4\") " pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:14 crc kubenswrapper[4690]: I1203 13:15:14.722525 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:15 crc kubenswrapper[4690]: I1203 13:15:15.169695 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w"] Dec 03 13:15:15 crc kubenswrapper[4690]: W1203 13:15:15.174920 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc080a74_f23e_4f89_9fc0_8cab2012a2f4.slice/crio-eaad3c784ef92016c0602b588ad4d7ec8e31387c7b3dcda954e6b16910e7f4f0 WatchSource:0}: Error finding container eaad3c784ef92016c0602b588ad4d7ec8e31387c7b3dcda954e6b16910e7f4f0: Status 404 returned error can't find the container with id eaad3c784ef92016c0602b588ad4d7ec8e31387c7b3dcda954e6b16910e7f4f0 Dec 03 13:15:15 crc kubenswrapper[4690]: I1203 13:15:15.194591 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" event={"ID":"ed128bdd-2a3f-4286-9ea8-295681e5cc89","Type":"ContainerDied","Data":"4fce32aef77fd75c5628bf2e21bbe065a45688d4856c6fb4186b180cce7e44d8"} Dec 03 13:15:15 crc kubenswrapper[4690]: I1203 13:15:15.194719 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z" Dec 03 13:15:15 crc kubenswrapper[4690]: I1203 13:15:15.195005 4690 scope.go:117] "RemoveContainer" containerID="19c8b9caa4f383145f796dcfc5efe8c121342842a393d5bb2ac404cf915b0f40" Dec 03 13:15:15 crc kubenswrapper[4690]: I1203 13:15:15.225934 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z"] Dec 03 13:15:15 crc kubenswrapper[4690]: I1203 13:15:15.229047 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-557bb68cfc-lpw6z"] Dec 03 13:15:16 crc kubenswrapper[4690]: I1203 13:15:16.203281 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" event={"ID":"cc080a74-f23e-4f89-9fc0-8cab2012a2f4","Type":"ContainerStarted","Data":"a1dad50cede75795e3585cdb23821a5fb74ab81c07298746584b816a0dbbb952"} Dec 03 13:15:16 crc kubenswrapper[4690]: I1203 13:15:16.203342 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" event={"ID":"cc080a74-f23e-4f89-9fc0-8cab2012a2f4","Type":"ContainerStarted","Data":"eaad3c784ef92016c0602b588ad4d7ec8e31387c7b3dcda954e6b16910e7f4f0"} Dec 03 13:15:16 crc kubenswrapper[4690]: I1203 13:15:16.203632 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:16 crc kubenswrapper[4690]: I1203 13:15:16.208598 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" Dec 03 13:15:16 crc kubenswrapper[4690]: I1203 13:15:16.225694 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-86cbd6df86-2rw5w" podStartSLOduration=4.225667938 podStartE2EDuration="4.225667938s" podCreationTimestamp="2025-12-03 13:15:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:15:16.224070584 +0000 UTC m=+362.204991037" watchObservedRunningTime="2025-12-03 13:15:16.225667938 +0000 UTC m=+362.206588371" Dec 03 13:15:16 crc kubenswrapper[4690]: I1203 13:15:16.322086 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed128bdd-2a3f-4286-9ea8-295681e5cc89" path="/var/lib/kubelet/pods/ed128bdd-2a3f-4286-9ea8-295681e5cc89/volumes" Dec 03 13:15:16 crc kubenswrapper[4690]: I1203 13:15:16.823540 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:15:16 crc kubenswrapper[4690]: I1203 13:15:16.824008 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.124174 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w9wd6"] Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.127471 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.131457 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.141790 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9wd6"] Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.312189 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vvnj\" (UniqueName: \"kubernetes.io/projected/d6a489f2-fe20-4d26-87f4-82c65ae7a461-kube-api-access-7vvnj\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.312267 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6a489f2-fe20-4d26-87f4-82c65ae7a461-catalog-content\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.312292 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6a489f2-fe20-4d26-87f4-82c65ae7a461-utilities\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.326998 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p4jzs"] Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.328315 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.330699 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.346795 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p4jzs"] Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.414032 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vvnj\" (UniqueName: \"kubernetes.io/projected/d6a489f2-fe20-4d26-87f4-82c65ae7a461-kube-api-access-7vvnj\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.414108 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6a489f2-fe20-4d26-87f4-82c65ae7a461-catalog-content\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.414144 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6a489f2-fe20-4d26-87f4-82c65ae7a461-utilities\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.414648 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6a489f2-fe20-4d26-87f4-82c65ae7a461-utilities\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.414664 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6a489f2-fe20-4d26-87f4-82c65ae7a461-catalog-content\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.438446 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vvnj\" (UniqueName: \"kubernetes.io/projected/d6a489f2-fe20-4d26-87f4-82c65ae7a461-kube-api-access-7vvnj\") pod \"redhat-marketplace-w9wd6\" (UID: \"d6a489f2-fe20-4d26-87f4-82c65ae7a461\") " pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.446958 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.524376 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-utilities\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.524462 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq7nh\" (UniqueName: \"kubernetes.io/projected/3d01a069-e06b-42e6-9667-0c8e57306a2a-kube-api-access-rq7nh\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.524561 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-catalog-content\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.625972 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-catalog-content\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.626481 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-utilities\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.626529 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq7nh\" (UniqueName: \"kubernetes.io/projected/3d01a069-e06b-42e6-9667-0c8e57306a2a-kube-api-access-rq7nh\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.626795 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-catalog-content\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.627158 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-utilities\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.648193 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq7nh\" (UniqueName: \"kubernetes.io/projected/3d01a069-e06b-42e6-9667-0c8e57306a2a-kube-api-access-rq7nh\") pod \"redhat-operators-p4jzs\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.856251 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9wd6"] Dec 03 13:15:21 crc kubenswrapper[4690]: I1203 13:15:21.942708 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:22 crc kubenswrapper[4690]: I1203 13:15:22.238819 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9wd6" event={"ID":"d6a489f2-fe20-4d26-87f4-82c65ae7a461","Type":"ContainerStarted","Data":"a59641bd801bffb7c764456f376b0fa5aff222fa6028f93c1a0a59cf632b842d"} Dec 03 13:15:22 crc kubenswrapper[4690]: I1203 13:15:22.336841 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p4jzs"] Dec 03 13:15:22 crc kubenswrapper[4690]: W1203 13:15:22.342644 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d01a069_e06b_42e6_9667_0c8e57306a2a.slice/crio-0915cf7e938e0fb8a9a55ad0b0f0e35e34befcb3c9c81bc7302e04d488438799 WatchSource:0}: Error finding container 0915cf7e938e0fb8a9a55ad0b0f0e35e34befcb3c9c81bc7302e04d488438799: Status 404 returned error can't find the container with id 0915cf7e938e0fb8a9a55ad0b0f0e35e34befcb3c9c81bc7302e04d488438799 Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.247520 4690 generic.go:334] "Generic (PLEG): container finished" podID="d6a489f2-fe20-4d26-87f4-82c65ae7a461" containerID="65cd8e0f595bcacf80741b8c618df1ee394c8dcaa8e00049751675068976170f" exitCode=0 Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.247610 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9wd6" event={"ID":"d6a489f2-fe20-4d26-87f4-82c65ae7a461","Type":"ContainerDied","Data":"65cd8e0f595bcacf80741b8c618df1ee394c8dcaa8e00049751675068976170f"} Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.251346 4690 generic.go:334] "Generic (PLEG): container finished" podID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerID="33ec16a8131c112a06881b898fd70a6d0821e42c45b976cc7a7684c605ccfb58" exitCode=0 Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.251388 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4jzs" event={"ID":"3d01a069-e06b-42e6-9667-0c8e57306a2a","Type":"ContainerDied","Data":"33ec16a8131c112a06881b898fd70a6d0821e42c45b976cc7a7684c605ccfb58"} Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.251417 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4jzs" event={"ID":"3d01a069-e06b-42e6-9667-0c8e57306a2a","Type":"ContainerStarted","Data":"0915cf7e938e0fb8a9a55ad0b0f0e35e34befcb3c9c81bc7302e04d488438799"} Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.723634 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z9fw6"] Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.724687 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.729258 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.739488 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z9fw6"] Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.857313 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/096d833c-6989-4a0e-8437-bdbc751a603d-catalog-content\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.857358 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/096d833c-6989-4a0e-8437-bdbc751a603d-utilities\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.857393 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msbpr\" (UniqueName: \"kubernetes.io/projected/096d833c-6989-4a0e-8437-bdbc751a603d-kube-api-access-msbpr\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.928757 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bbjwm"] Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.929755 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.932162 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.942217 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bbjwm"] Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.960554 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a384d66-783b-45bc-afa5-fb9d77e223ac-catalog-content\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.960607 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a384d66-783b-45bc-afa5-fb9d77e223ac-utilities\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.960655 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvzvp\" (UniqueName: \"kubernetes.io/projected/4a384d66-783b-45bc-afa5-fb9d77e223ac-kube-api-access-xvzvp\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.960696 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/096d833c-6989-4a0e-8437-bdbc751a603d-catalog-content\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.960730 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/096d833c-6989-4a0e-8437-bdbc751a603d-utilities\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.960780 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msbpr\" (UniqueName: \"kubernetes.io/projected/096d833c-6989-4a0e-8437-bdbc751a603d-kube-api-access-msbpr\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.962097 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/096d833c-6989-4a0e-8437-bdbc751a603d-catalog-content\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.962269 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/096d833c-6989-4a0e-8437-bdbc751a603d-utilities\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:23 crc kubenswrapper[4690]: I1203 13:15:23.982023 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msbpr\" (UniqueName: \"kubernetes.io/projected/096d833c-6989-4a0e-8437-bdbc751a603d-kube-api-access-msbpr\") pod \"community-operators-z9fw6\" (UID: \"096d833c-6989-4a0e-8437-bdbc751a603d\") " pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.046421 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.062131 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a384d66-783b-45bc-afa5-fb9d77e223ac-catalog-content\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.062524 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a384d66-783b-45bc-afa5-fb9d77e223ac-utilities\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.062555 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvzvp\" (UniqueName: \"kubernetes.io/projected/4a384d66-783b-45bc-afa5-fb9d77e223ac-kube-api-access-xvzvp\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.062631 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a384d66-783b-45bc-afa5-fb9d77e223ac-catalog-content\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.062971 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a384d66-783b-45bc-afa5-fb9d77e223ac-utilities\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.078585 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvzvp\" (UniqueName: \"kubernetes.io/projected/4a384d66-783b-45bc-afa5-fb9d77e223ac-kube-api-access-xvzvp\") pod \"certified-operators-bbjwm\" (UID: \"4a384d66-783b-45bc-afa5-fb9d77e223ac\") " pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.244338 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.436633 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z9fw6"] Dec 03 13:15:24 crc kubenswrapper[4690]: W1203 13:15:24.439603 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod096d833c_6989_4a0e_8437_bdbc751a603d.slice/crio-9591df606c24a21494eff767caa43884105bf50cdf6db5f90004e51d2130c157 WatchSource:0}: Error finding container 9591df606c24a21494eff767caa43884105bf50cdf6db5f90004e51d2130c157: Status 404 returned error can't find the container with id 9591df606c24a21494eff767caa43884105bf50cdf6db5f90004e51d2130c157 Dec 03 13:15:24 crc kubenswrapper[4690]: I1203 13:15:24.650986 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bbjwm"] Dec 03 13:15:24 crc kubenswrapper[4690]: W1203 13:15:24.655482 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a384d66_783b_45bc_afa5_fb9d77e223ac.slice/crio-12dc875c9aa5d0599c0b2a68472f0cff0e1d1301ef437637112c2c25510f7333 WatchSource:0}: Error finding container 12dc875c9aa5d0599c0b2a68472f0cff0e1d1301ef437637112c2c25510f7333: Status 404 returned error can't find the container with id 12dc875c9aa5d0599c0b2a68472f0cff0e1d1301ef437637112c2c25510f7333 Dec 03 13:15:25 crc kubenswrapper[4690]: I1203 13:15:25.264346 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbjwm" event={"ID":"4a384d66-783b-45bc-afa5-fb9d77e223ac","Type":"ContainerStarted","Data":"12dc875c9aa5d0599c0b2a68472f0cff0e1d1301ef437637112c2c25510f7333"} Dec 03 13:15:25 crc kubenswrapper[4690]: I1203 13:15:25.265551 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9fw6" event={"ID":"096d833c-6989-4a0e-8437-bdbc751a603d","Type":"ContainerStarted","Data":"35863c4369436263ff057e5a8cd271019c2196a47eb8c3535844e59aa46ee052"} Dec 03 13:15:25 crc kubenswrapper[4690]: I1203 13:15:25.265582 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9fw6" event={"ID":"096d833c-6989-4a0e-8437-bdbc751a603d","Type":"ContainerStarted","Data":"9591df606c24a21494eff767caa43884105bf50cdf6db5f90004e51d2130c157"} Dec 03 13:15:26 crc kubenswrapper[4690]: I1203 13:15:26.272191 4690 generic.go:334] "Generic (PLEG): container finished" podID="4a384d66-783b-45bc-afa5-fb9d77e223ac" containerID="e6b3e4e50db6b2d1c9b7a1a4d7b70ad03909dc6a1dfc6b52c1d93f129c477aa2" exitCode=0 Dec 03 13:15:26 crc kubenswrapper[4690]: I1203 13:15:26.272397 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbjwm" event={"ID":"4a384d66-783b-45bc-afa5-fb9d77e223ac","Type":"ContainerDied","Data":"e6b3e4e50db6b2d1c9b7a1a4d7b70ad03909dc6a1dfc6b52c1d93f129c477aa2"} Dec 03 13:15:26 crc kubenswrapper[4690]: I1203 13:15:26.274995 4690 generic.go:334] "Generic (PLEG): container finished" podID="096d833c-6989-4a0e-8437-bdbc751a603d" containerID="35863c4369436263ff057e5a8cd271019c2196a47eb8c3535844e59aa46ee052" exitCode=0 Dec 03 13:15:26 crc kubenswrapper[4690]: I1203 13:15:26.275038 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9fw6" event={"ID":"096d833c-6989-4a0e-8437-bdbc751a603d","Type":"ContainerDied","Data":"35863c4369436263ff057e5a8cd271019c2196a47eb8c3535844e59aa46ee052"} Dec 03 13:15:29 crc kubenswrapper[4690]: I1203 13:15:29.296428 4690 generic.go:334] "Generic (PLEG): container finished" podID="d6a489f2-fe20-4d26-87f4-82c65ae7a461" containerID="495b98fc19c9820ccefcae697418e21832fda47e58b49824e802c2b72f76b176" exitCode=0 Dec 03 13:15:29 crc kubenswrapper[4690]: I1203 13:15:29.296523 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9wd6" event={"ID":"d6a489f2-fe20-4d26-87f4-82c65ae7a461","Type":"ContainerDied","Data":"495b98fc19c9820ccefcae697418e21832fda47e58b49824e802c2b72f76b176"} Dec 03 13:15:29 crc kubenswrapper[4690]: I1203 13:15:29.299512 4690 generic.go:334] "Generic (PLEG): container finished" podID="4a384d66-783b-45bc-afa5-fb9d77e223ac" containerID="7e188e79288e17d05a0f513fef73ac52d880f9ad5ed9505eba861e5461f06c43" exitCode=0 Dec 03 13:15:29 crc kubenswrapper[4690]: I1203 13:15:29.299591 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbjwm" event={"ID":"4a384d66-783b-45bc-afa5-fb9d77e223ac","Type":"ContainerDied","Data":"7e188e79288e17d05a0f513fef73ac52d880f9ad5ed9505eba861e5461f06c43"} Dec 03 13:15:29 crc kubenswrapper[4690]: I1203 13:15:29.301665 4690 generic.go:334] "Generic (PLEG): container finished" podID="096d833c-6989-4a0e-8437-bdbc751a603d" containerID="0876f24a0494ebeaf9db3df303eabc95bfb2521095634d98ff593560d7f92341" exitCode=0 Dec 03 13:15:29 crc kubenswrapper[4690]: I1203 13:15:29.302235 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9fw6" event={"ID":"096d833c-6989-4a0e-8437-bdbc751a603d","Type":"ContainerDied","Data":"0876f24a0494ebeaf9db3df303eabc95bfb2521095634d98ff593560d7f92341"} Dec 03 13:15:29 crc kubenswrapper[4690]: I1203 13:15:29.305080 4690 generic.go:334] "Generic (PLEG): container finished" podID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerID="ae5c1135e3570e449535946c2f1abb29081239f608c8b454c1a2efa92ca907a5" exitCode=0 Dec 03 13:15:29 crc kubenswrapper[4690]: I1203 13:15:29.305157 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4jzs" event={"ID":"3d01a069-e06b-42e6-9667-0c8e57306a2a","Type":"ContainerDied","Data":"ae5c1135e3570e449535946c2f1abb29081239f608c8b454c1a2efa92ca907a5"} Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.311569 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbjwm" event={"ID":"4a384d66-783b-45bc-afa5-fb9d77e223ac","Type":"ContainerStarted","Data":"9f05db004c030eb6f19908a78f45a1aaa749e415cc1c18fe423521634f753868"} Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.320501 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9fw6" event={"ID":"096d833c-6989-4a0e-8437-bdbc751a603d","Type":"ContainerStarted","Data":"5b32b6e375c41b5168476ec8599470a1c7fcb32bc801fdec29cf245f596da3cc"} Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.321211 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4jzs" event={"ID":"3d01a069-e06b-42e6-9667-0c8e57306a2a","Type":"ContainerStarted","Data":"81d4d3c516acfacedd8298b5b036aa71df6234922824d3a4819d535cbbbad77c"} Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.322962 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9wd6" event={"ID":"d6a489f2-fe20-4d26-87f4-82c65ae7a461","Type":"ContainerStarted","Data":"29b05b20f8f864235a74e09c55a95841eee4605fe7857ee32ac0f5c07767e500"} Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.340785 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bbjwm" podStartSLOduration=3.978414538 podStartE2EDuration="7.34076538s" podCreationTimestamp="2025-12-03 13:15:23 +0000 UTC" firstStartedPulling="2025-12-03 13:15:26.430579829 +0000 UTC m=+372.411500262" lastFinishedPulling="2025-12-03 13:15:29.792930671 +0000 UTC m=+375.773851104" observedRunningTime="2025-12-03 13:15:30.339225507 +0000 UTC m=+376.320145950" watchObservedRunningTime="2025-12-03 13:15:30.34076538 +0000 UTC m=+376.321685813" Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.351047 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-2l4wj" Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.361533 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z9fw6" podStartSLOduration=3.986989961 podStartE2EDuration="7.361492614s" podCreationTimestamp="2025-12-03 13:15:23 +0000 UTC" firstStartedPulling="2025-12-03 13:15:26.430570008 +0000 UTC m=+372.411490441" lastFinishedPulling="2025-12-03 13:15:29.805072661 +0000 UTC m=+375.785993094" observedRunningTime="2025-12-03 13:15:30.359048732 +0000 UTC m=+376.339969155" watchObservedRunningTime="2025-12-03 13:15:30.361492614 +0000 UTC m=+376.342413047" Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.384116 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p4jzs" podStartSLOduration=2.920690634 podStartE2EDuration="9.384087518s" podCreationTimestamp="2025-12-03 13:15:21 +0000 UTC" firstStartedPulling="2025-12-03 13:15:23.25363124 +0000 UTC m=+369.234551673" lastFinishedPulling="2025-12-03 13:15:29.717028124 +0000 UTC m=+375.697948557" observedRunningTime="2025-12-03 13:15:30.380383759 +0000 UTC m=+376.361304192" watchObservedRunningTime="2025-12-03 13:15:30.384087518 +0000 UTC m=+376.365007951" Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.429288 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w9wd6" podStartSLOduration=2.993156946 podStartE2EDuration="9.429258386s" podCreationTimestamp="2025-12-03 13:15:21 +0000 UTC" firstStartedPulling="2025-12-03 13:15:23.249956691 +0000 UTC m=+369.230877124" lastFinishedPulling="2025-12-03 13:15:29.686058121 +0000 UTC m=+375.666978564" observedRunningTime="2025-12-03 13:15:30.404890204 +0000 UTC m=+376.385810657" watchObservedRunningTime="2025-12-03 13:15:30.429258386 +0000 UTC m=+376.410178819" Dec 03 13:15:30 crc kubenswrapper[4690]: I1203 13:15:30.480879 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lds4l"] Dec 03 13:15:31 crc kubenswrapper[4690]: I1203 13:15:31.448771 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:31 crc kubenswrapper[4690]: I1203 13:15:31.449241 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:31 crc kubenswrapper[4690]: I1203 13:15:31.498972 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:31 crc kubenswrapper[4690]: I1203 13:15:31.943480 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:31 crc kubenswrapper[4690]: I1203 13:15:31.943683 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.368998 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c"] Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.369270 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" podUID="276f20a9-9241-4ea1-a33f-7def78f05efd" containerName="controller-manager" containerID="cri-o://0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5" gracePeriod=30 Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.786925 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.885675 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-client-ca\") pod \"276f20a9-9241-4ea1-a33f-7def78f05efd\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.885758 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-proxy-ca-bundles\") pod \"276f20a9-9241-4ea1-a33f-7def78f05efd\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.885790 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69fg6\" (UniqueName: \"kubernetes.io/projected/276f20a9-9241-4ea1-a33f-7def78f05efd-kube-api-access-69fg6\") pod \"276f20a9-9241-4ea1-a33f-7def78f05efd\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.886054 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276f20a9-9241-4ea1-a33f-7def78f05efd-serving-cert\") pod \"276f20a9-9241-4ea1-a33f-7def78f05efd\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.886766 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "276f20a9-9241-4ea1-a33f-7def78f05efd" (UID: "276f20a9-9241-4ea1-a33f-7def78f05efd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.886801 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-client-ca" (OuterVolumeSpecName: "client-ca") pod "276f20a9-9241-4ea1-a33f-7def78f05efd" (UID: "276f20a9-9241-4ea1-a33f-7def78f05efd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.886980 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-config\") pod \"276f20a9-9241-4ea1-a33f-7def78f05efd\" (UID: \"276f20a9-9241-4ea1-a33f-7def78f05efd\") " Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.887436 4690 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.887454 4690 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.888045 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-config" (OuterVolumeSpecName: "config") pod "276f20a9-9241-4ea1-a33f-7def78f05efd" (UID: "276f20a9-9241-4ea1-a33f-7def78f05efd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.895438 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/276f20a9-9241-4ea1-a33f-7def78f05efd-kube-api-access-69fg6" (OuterVolumeSpecName: "kube-api-access-69fg6") pod "276f20a9-9241-4ea1-a33f-7def78f05efd" (UID: "276f20a9-9241-4ea1-a33f-7def78f05efd"). InnerVolumeSpecName "kube-api-access-69fg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.907944 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/276f20a9-9241-4ea1-a33f-7def78f05efd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "276f20a9-9241-4ea1-a33f-7def78f05efd" (UID: "276f20a9-9241-4ea1-a33f-7def78f05efd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.982634 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p4jzs" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="registry-server" probeResult="failure" output=< Dec 03 13:15:32 crc kubenswrapper[4690]: timeout: failed to connect service ":50051" within 1s Dec 03 13:15:32 crc kubenswrapper[4690]: > Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.988655 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69fg6\" (UniqueName: \"kubernetes.io/projected/276f20a9-9241-4ea1-a33f-7def78f05efd-kube-api-access-69fg6\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.988686 4690 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/276f20a9-9241-4ea1-a33f-7def78f05efd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:32 crc kubenswrapper[4690]: I1203 13:15:32.988698 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276f20a9-9241-4ea1-a33f-7def78f05efd-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.341171 4690 generic.go:334] "Generic (PLEG): container finished" podID="276f20a9-9241-4ea1-a33f-7def78f05efd" containerID="0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5" exitCode=0 Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.341219 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" event={"ID":"276f20a9-9241-4ea1-a33f-7def78f05efd","Type":"ContainerDied","Data":"0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5"} Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.341256 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.341288 4690 scope.go:117] "RemoveContainer" containerID="0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.341273 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c" event={"ID":"276f20a9-9241-4ea1-a33f-7def78f05efd","Type":"ContainerDied","Data":"87effcec771a21ffdbf33a58226d32ba4f237de8e0c3487427fd16238c9c5c11"} Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.356888 4690 scope.go:117] "RemoveContainer" containerID="0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5" Dec 03 13:15:33 crc kubenswrapper[4690]: E1203 13:15:33.357666 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5\": container with ID starting with 0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5 not found: ID does not exist" containerID="0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.357746 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5"} err="failed to get container status \"0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5\": rpc error: code = NotFound desc = could not find container \"0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5\": container with ID starting with 0c9e41fe0f0f131dce99b1caadf7f20544eaced7e0629dd8852ee65c3d339df5 not found: ID does not exist" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.368994 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c"] Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.373508 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-ltm7c"] Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.511110 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-8d4d5446-wztjj"] Dec 03 13:15:33 crc kubenswrapper[4690]: E1203 13:15:33.511386 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276f20a9-9241-4ea1-a33f-7def78f05efd" containerName="controller-manager" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.511403 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="276f20a9-9241-4ea1-a33f-7def78f05efd" containerName="controller-manager" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.511540 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="276f20a9-9241-4ea1-a33f-7def78f05efd" containerName="controller-manager" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.512017 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.517203 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.517462 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.518299 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.518724 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.519110 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.519324 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.523325 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8d4d5446-wztjj"] Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.528578 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.595258 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025762a6-f05f-4060-9b84-05d632a5172a-serving-cert\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.595312 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-client-ca\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.595449 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-proxy-ca-bundles\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.595486 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-config\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.595506 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tm7z\" (UniqueName: \"kubernetes.io/projected/025762a6-f05f-4060-9b84-05d632a5172a-kube-api-access-6tm7z\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.696357 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-proxy-ca-bundles\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.696437 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-config\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.696466 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tm7z\" (UniqueName: \"kubernetes.io/projected/025762a6-f05f-4060-9b84-05d632a5172a-kube-api-access-6tm7z\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.696547 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025762a6-f05f-4060-9b84-05d632a5172a-serving-cert\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.696574 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-client-ca\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.697936 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-proxy-ca-bundles\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.698031 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-client-ca\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.698175 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025762a6-f05f-4060-9b84-05d632a5172a-config\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.700893 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025762a6-f05f-4060-9b84-05d632a5172a-serving-cert\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.712719 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tm7z\" (UniqueName: \"kubernetes.io/projected/025762a6-f05f-4060-9b84-05d632a5172a-kube-api-access-6tm7z\") pod \"controller-manager-8d4d5446-wztjj\" (UID: \"025762a6-f05f-4060-9b84-05d632a5172a\") " pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:33 crc kubenswrapper[4690]: I1203 13:15:33.834682 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.046835 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.047317 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.085687 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8d4d5446-wztjj"] Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.103565 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.245049 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.246371 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.285228 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.322481 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="276f20a9-9241-4ea1-a33f-7def78f05efd" path="/var/lib/kubelet/pods/276f20a9-9241-4ea1-a33f-7def78f05efd/volumes" Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.349659 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" event={"ID":"025762a6-f05f-4060-9b84-05d632a5172a","Type":"ContainerStarted","Data":"7891f7640bd0ab72e081f0f9487c737f71bb8fb8c7dfb44a074f414f8155f52f"} Dec 03 13:15:34 crc kubenswrapper[4690]: I1203 13:15:34.384244 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z9fw6" Dec 03 13:15:35 crc kubenswrapper[4690]: I1203 13:15:35.356195 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" event={"ID":"025762a6-f05f-4060-9b84-05d632a5172a","Type":"ContainerStarted","Data":"441bfc526beb9d7d3469657f34993235c7b79d1f443cdadeff7a046b67ca0878"} Dec 03 13:15:35 crc kubenswrapper[4690]: I1203 13:15:35.392224 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bbjwm" Dec 03 13:15:36 crc kubenswrapper[4690]: I1203 13:15:36.388758 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" podStartSLOduration=4.388734562 podStartE2EDuration="4.388734562s" podCreationTimestamp="2025-12-03 13:15:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:15:36.385086084 +0000 UTC m=+382.366006527" watchObservedRunningTime="2025-12-03 13:15:36.388734562 +0000 UTC m=+382.369655005" Dec 03 13:15:41 crc kubenswrapper[4690]: I1203 13:15:41.496485 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w9wd6" Dec 03 13:15:41 crc kubenswrapper[4690]: I1203 13:15:41.986639 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:42 crc kubenswrapper[4690]: I1203 13:15:42.030376 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 13:15:43 crc kubenswrapper[4690]: I1203 13:15:43.835101 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:43 crc kubenswrapper[4690]: I1203 13:15:43.841000 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-8d4d5446-wztjj" Dec 03 13:15:46 crc kubenswrapper[4690]: I1203 13:15:46.823748 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:15:46 crc kubenswrapper[4690]: I1203 13:15:46.824202 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:15:55 crc kubenswrapper[4690]: I1203 13:15:55.527131 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" podUID="3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" containerName="registry" containerID="cri-o://f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9" gracePeriod=30 Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.143741 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.273491 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-tls\") pod \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.273542 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-installation-pull-secrets\") pod \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.273564 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-bound-sa-token\") pod \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.273762 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.273816 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-trusted-ca\") pod \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.273842 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6lvq\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-kube-api-access-t6lvq\") pod \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.273933 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-certificates\") pod \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.273949 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-ca-trust-extracted\") pod \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\" (UID: \"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a\") " Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.275641 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.275932 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.280417 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.280546 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.280866 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-kube-api-access-t6lvq" (OuterVolumeSpecName: "kube-api-access-t6lvq") pod "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a"). InnerVolumeSpecName "kube-api-access-t6lvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.281217 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.285797 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 13:15:56 crc kubenswrapper[4690]: I1203 13:15:56.290939 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" (UID: "3b8dc7b6-5e28-49c8-a0d9-4c17340d579a"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.376060 4690 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.376098 4690 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.376113 4690 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.376123 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.376135 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6lvq\" (UniqueName: \"kubernetes.io/projected/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-kube-api-access-t6lvq\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.376146 4690 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.376157 4690 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.487756 4690 generic.go:334] "Generic (PLEG): container finished" podID="3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" containerID="f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9" exitCode=0 Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.487797 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" event={"ID":"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a","Type":"ContainerDied","Data":"f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9"} Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.487830 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" event={"ID":"3b8dc7b6-5e28-49c8-a0d9-4c17340d579a","Type":"ContainerDied","Data":"88806e92948e40dabb06baa0be3e96958d30ed1fd79476a114e1941c67d289b9"} Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.487859 4690 scope.go:117] "RemoveContainer" containerID="f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.487855 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lds4l" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.512476 4690 scope.go:117] "RemoveContainer" containerID="f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9" Dec 03 13:15:58 crc kubenswrapper[4690]: E1203 13:15:56.514818 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9\": container with ID starting with f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9 not found: ID does not exist" containerID="f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.514846 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9"} err="failed to get container status \"f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9\": rpc error: code = NotFound desc = could not find container \"f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9\": container with ID starting with f690d5dfa8d7e7ba6f741ebddbcdd2c11d2e5e1dc4a12fe11bb52d1e3d1de2d9 not found: ID does not exist" Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.514907 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lds4l"] Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:56.518714 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lds4l"] Dec 03 13:15:58 crc kubenswrapper[4690]: I1203 13:15:58.321298 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" path="/var/lib/kubelet/pods/3b8dc7b6-5e28-49c8-a0d9-4c17340d579a/volumes" Dec 03 13:16:16 crc kubenswrapper[4690]: I1203 13:16:16.823758 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:16:16 crc kubenswrapper[4690]: I1203 13:16:16.824649 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:16:16 crc kubenswrapper[4690]: I1203 13:16:16.824725 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:16:16 crc kubenswrapper[4690]: I1203 13:16:16.825605 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9e92c52ce940ca7b19ac9c6c34c60bff5ccafb59832fe7eb80f99d62ea7a9e8"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:16:16 crc kubenswrapper[4690]: I1203 13:16:16.825674 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://c9e92c52ce940ca7b19ac9c6c34c60bff5ccafb59832fe7eb80f99d62ea7a9e8" gracePeriod=600 Dec 03 13:16:17 crc kubenswrapper[4690]: I1203 13:16:17.614541 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="c9e92c52ce940ca7b19ac9c6c34c60bff5ccafb59832fe7eb80f99d62ea7a9e8" exitCode=0 Dec 03 13:16:17 crc kubenswrapper[4690]: I1203 13:16:17.614629 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"c9e92c52ce940ca7b19ac9c6c34c60bff5ccafb59832fe7eb80f99d62ea7a9e8"} Dec 03 13:16:17 crc kubenswrapper[4690]: I1203 13:16:17.615238 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"c69bb0d7919080bb99202ef400a31cbfc36de1e7353348ec9eea0dfce4422b6f"} Dec 03 13:16:17 crc kubenswrapper[4690]: I1203 13:16:17.615257 4690 scope.go:117] "RemoveContainer" containerID="cf623a03d002ae35f7d0d1a0e4c35574458b71de5445f79095675a46a24b2bed" Dec 03 13:18:46 crc kubenswrapper[4690]: I1203 13:18:46.823610 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:18:46 crc kubenswrapper[4690]: I1203 13:18:46.824295 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:19:16 crc kubenswrapper[4690]: I1203 13:19:16.823465 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:19:16 crc kubenswrapper[4690]: I1203 13:19:16.824072 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:19:46 crc kubenswrapper[4690]: I1203 13:19:46.824177 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:19:46 crc kubenswrapper[4690]: I1203 13:19:46.826185 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:19:46 crc kubenswrapper[4690]: I1203 13:19:46.826309 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:19:46 crc kubenswrapper[4690]: I1203 13:19:46.827518 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c69bb0d7919080bb99202ef400a31cbfc36de1e7353348ec9eea0dfce4422b6f"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:19:46 crc kubenswrapper[4690]: I1203 13:19:46.827674 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://c69bb0d7919080bb99202ef400a31cbfc36de1e7353348ec9eea0dfce4422b6f" gracePeriod=600 Dec 03 13:19:47 crc kubenswrapper[4690]: I1203 13:19:47.806073 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="c69bb0d7919080bb99202ef400a31cbfc36de1e7353348ec9eea0dfce4422b6f" exitCode=0 Dec 03 13:19:47 crc kubenswrapper[4690]: I1203 13:19:47.806150 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"c69bb0d7919080bb99202ef400a31cbfc36de1e7353348ec9eea0dfce4422b6f"} Dec 03 13:19:47 crc kubenswrapper[4690]: I1203 13:19:47.806686 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"6757d3773013b08b3fe1dc029582fe576a1dfe2c310afb7b61e2a8c64d0aa40f"} Dec 03 13:19:47 crc kubenswrapper[4690]: I1203 13:19:47.806730 4690 scope.go:117] "RemoveContainer" containerID="c9e92c52ce940ca7b19ac9c6c34c60bff5ccafb59832fe7eb80f99d62ea7a9e8" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.286672 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2p8k2"] Dec 03 13:20:37 crc kubenswrapper[4690]: E1203 13:20:37.287665 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" containerName="registry" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.287690 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" containerName="registry" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.287914 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8dc7b6-5e28-49c8-a0d9-4c17340d579a" containerName="registry" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.288565 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-2p8k2" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.293510 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.293756 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.293963 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2gp2s"] Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.294675 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2gp2s" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.297667 4690 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-qnphx" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.297757 4690 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-4qlsb" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.305831 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2gp2s"] Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.310379 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4jfnl"] Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.311284 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.317425 4690 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-2mpgn" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.318158 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2p8k2"] Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.335754 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4jfnl"] Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.410311 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sktt\" (UniqueName: \"kubernetes.io/projected/6daa0e47-6f3f-4ac8-b1a8-596401c33ef2-kube-api-access-5sktt\") pod \"cert-manager-cainjector-7f985d654d-2p8k2\" (UID: \"6daa0e47-6f3f-4ac8-b1a8-596401c33ef2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2p8k2" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.410418 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6tcq\" (UniqueName: \"kubernetes.io/projected/9c22c662-9a32-4721-9958-12651fa443a7-kube-api-access-x6tcq\") pod \"cert-manager-5b446d88c5-2gp2s\" (UID: \"9c22c662-9a32-4721-9958-12651fa443a7\") " pod="cert-manager/cert-manager-5b446d88c5-2gp2s" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.511556 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sktt\" (UniqueName: \"kubernetes.io/projected/6daa0e47-6f3f-4ac8-b1a8-596401c33ef2-kube-api-access-5sktt\") pod \"cert-manager-cainjector-7f985d654d-2p8k2\" (UID: \"6daa0e47-6f3f-4ac8-b1a8-596401c33ef2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2p8k2" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.511700 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6tcq\" (UniqueName: \"kubernetes.io/projected/9c22c662-9a32-4721-9958-12651fa443a7-kube-api-access-x6tcq\") pod \"cert-manager-5b446d88c5-2gp2s\" (UID: \"9c22c662-9a32-4721-9958-12651fa443a7\") " pod="cert-manager/cert-manager-5b446d88c5-2gp2s" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.511745 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqcsz\" (UniqueName: \"kubernetes.io/projected/f47a381d-4adf-4543-b4f3-793cdc3db649-kube-api-access-zqcsz\") pod \"cert-manager-webhook-5655c58dd6-4jfnl\" (UID: \"f47a381d-4adf-4543-b4f3-793cdc3db649\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.543162 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6tcq\" (UniqueName: \"kubernetes.io/projected/9c22c662-9a32-4721-9958-12651fa443a7-kube-api-access-x6tcq\") pod \"cert-manager-5b446d88c5-2gp2s\" (UID: \"9c22c662-9a32-4721-9958-12651fa443a7\") " pod="cert-manager/cert-manager-5b446d88c5-2gp2s" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.544949 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sktt\" (UniqueName: \"kubernetes.io/projected/6daa0e47-6f3f-4ac8-b1a8-596401c33ef2-kube-api-access-5sktt\") pod \"cert-manager-cainjector-7f985d654d-2p8k2\" (UID: \"6daa0e47-6f3f-4ac8-b1a8-596401c33ef2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2p8k2" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.609921 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-2p8k2" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.612583 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqcsz\" (UniqueName: \"kubernetes.io/projected/f47a381d-4adf-4543-b4f3-793cdc3db649-kube-api-access-zqcsz\") pod \"cert-manager-webhook-5655c58dd6-4jfnl\" (UID: \"f47a381d-4adf-4543-b4f3-793cdc3db649\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.626244 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2gp2s" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.632565 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqcsz\" (UniqueName: \"kubernetes.io/projected/f47a381d-4adf-4543-b4f3-793cdc3db649-kube-api-access-zqcsz\") pod \"cert-manager-webhook-5655c58dd6-4jfnl\" (UID: \"f47a381d-4adf-4543-b4f3-793cdc3db649\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.639948 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.841980 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2p8k2"] Dec 03 13:20:37 crc kubenswrapper[4690]: I1203 13:20:37.854369 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:20:38 crc kubenswrapper[4690]: I1203 13:20:38.107724 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4jfnl"] Dec 03 13:20:38 crc kubenswrapper[4690]: I1203 13:20:38.112185 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2gp2s"] Dec 03 13:20:38 crc kubenswrapper[4690]: W1203 13:20:38.123652 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c22c662_9a32_4721_9958_12651fa443a7.slice/crio-3b2e17410a3d847f8bf9581042352a46540cf4eaf1ac71d7543c50e43d1355be WatchSource:0}: Error finding container 3b2e17410a3d847f8bf9581042352a46540cf4eaf1ac71d7543c50e43d1355be: Status 404 returned error can't find the container with id 3b2e17410a3d847f8bf9581042352a46540cf4eaf1ac71d7543c50e43d1355be Dec 03 13:20:38 crc kubenswrapper[4690]: I1203 13:20:38.132676 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" event={"ID":"f47a381d-4adf-4543-b4f3-793cdc3db649","Type":"ContainerStarted","Data":"1408b5509867d934eeb1f2fd692ff375c1321c20d8abb69ff759a3938053ef3e"} Dec 03 13:20:38 crc kubenswrapper[4690]: I1203 13:20:38.134221 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-2p8k2" event={"ID":"6daa0e47-6f3f-4ac8-b1a8-596401c33ef2","Type":"ContainerStarted","Data":"92b44c4399a410b2e9cc04df466f31578fec0dff4059ef8e83f48ca850042384"} Dec 03 13:20:38 crc kubenswrapper[4690]: I1203 13:20:38.135138 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2gp2s" event={"ID":"9c22c662-9a32-4721-9958-12651fa443a7","Type":"ContainerStarted","Data":"3b2e17410a3d847f8bf9581042352a46540cf4eaf1ac71d7543c50e43d1355be"} Dec 03 13:20:42 crc kubenswrapper[4690]: I1203 13:20:42.163742 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" event={"ID":"f47a381d-4adf-4543-b4f3-793cdc3db649","Type":"ContainerStarted","Data":"3d3c4e73d7f3177a79301c8974d8677e5041fef05670e6659cc7458131fcbc54"} Dec 03 13:20:42 crc kubenswrapper[4690]: I1203 13:20:42.164420 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" Dec 03 13:20:42 crc kubenswrapper[4690]: I1203 13:20:42.166234 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-2p8k2" event={"ID":"6daa0e47-6f3f-4ac8-b1a8-596401c33ef2","Type":"ContainerStarted","Data":"19c08eaf6e9182935975c5085cf9f1dd69c633c4deb9e002c28dfccae4768eb6"} Dec 03 13:20:42 crc kubenswrapper[4690]: I1203 13:20:42.168478 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2gp2s" event={"ID":"9c22c662-9a32-4721-9958-12651fa443a7","Type":"ContainerStarted","Data":"9cb946a5c4468d2a2cb1b77197f42a15c8bf4e3320013cc466fca1b7b0f4a43a"} Dec 03 13:20:42 crc kubenswrapper[4690]: I1203 13:20:42.181524 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" podStartSLOduration=2.313714928 podStartE2EDuration="5.181502531s" podCreationTimestamp="2025-12-03 13:20:37 +0000 UTC" firstStartedPulling="2025-12-03 13:20:38.112589534 +0000 UTC m=+684.093510007" lastFinishedPulling="2025-12-03 13:20:40.980377167 +0000 UTC m=+686.961297610" observedRunningTime="2025-12-03 13:20:42.180464044 +0000 UTC m=+688.161384497" watchObservedRunningTime="2025-12-03 13:20:42.181502531 +0000 UTC m=+688.162422964" Dec 03 13:20:42 crc kubenswrapper[4690]: I1203 13:20:42.197853 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-2p8k2" podStartSLOduration=2.073410167 podStartE2EDuration="5.197836023s" podCreationTimestamp="2025-12-03 13:20:37 +0000 UTC" firstStartedPulling="2025-12-03 13:20:37.854154064 +0000 UTC m=+683.835074497" lastFinishedPulling="2025-12-03 13:20:40.97857992 +0000 UTC m=+686.959500353" observedRunningTime="2025-12-03 13:20:42.195146654 +0000 UTC m=+688.176067107" watchObservedRunningTime="2025-12-03 13:20:42.197836023 +0000 UTC m=+688.178756456" Dec 03 13:20:42 crc kubenswrapper[4690]: I1203 13:20:42.220048 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-2gp2s" podStartSLOduration=2.305164816 podStartE2EDuration="5.220021986s" podCreationTimestamp="2025-12-03 13:20:37 +0000 UTC" firstStartedPulling="2025-12-03 13:20:38.126421161 +0000 UTC m=+684.107341594" lastFinishedPulling="2025-12-03 13:20:41.041278331 +0000 UTC m=+687.022198764" observedRunningTime="2025-12-03 13:20:42.215577872 +0000 UTC m=+688.196498345" watchObservedRunningTime="2025-12-03 13:20:42.220021986 +0000 UTC m=+688.200942429" Dec 03 13:20:47 crc kubenswrapper[4690]: I1203 13:20:47.644053 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-4jfnl" Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.307839 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vmddn"] Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.308887 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovn-controller" containerID="cri-o://4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf" gracePeriod=30 Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.309418 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="sbdb" containerID="cri-o://8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" gracePeriod=30 Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.309464 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="nbdb" containerID="cri-o://149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" gracePeriod=30 Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.309496 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="northd" containerID="cri-o://79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3" gracePeriod=30 Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.309530 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555" gracePeriod=30 Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.309567 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kube-rbac-proxy-node" containerID="cri-o://53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a" gracePeriod=30 Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.309643 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovn-acl-logging" containerID="cri-o://fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628" gracePeriod=30 Dec 03 13:21:04 crc kubenswrapper[4690]: I1203 13:21:04.352597 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" containerID="cri-o://950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" gracePeriod=30 Dec 03 13:21:04 crc kubenswrapper[4690]: E1203 13:21:04.494778 4690 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 03 13:21:04 crc kubenswrapper[4690]: E1203 13:21:04.494850 4690 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 03 13:21:04 crc kubenswrapper[4690]: E1203 13:21:04.496439 4690 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 03 13:21:04 crc kubenswrapper[4690]: E1203 13:21:04.496708 4690 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 03 13:21:04 crc kubenswrapper[4690]: E1203 13:21:04.497733 4690 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 03 13:21:04 crc kubenswrapper[4690]: E1203 13:21:04.497782 4690 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="nbdb" Dec 03 13:21:04 crc kubenswrapper[4690]: E1203 13:21:04.498489 4690 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 03 13:21:04 crc kubenswrapper[4690]: E1203 13:21:04.498535 4690 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="sbdb" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.109900 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/3.log" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.112727 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovn-acl-logging/0.log" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.113247 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovn-controller/0.log" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.113672 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.178519 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2h5pp"] Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.178811 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.178917 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.178949 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kube-rbac-proxy-node" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.178956 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kube-rbac-proxy-node" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.178970 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.178977 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.178984 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.178991 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179001 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kubecfg-setup" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179029 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kubecfg-setup" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179040 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179045 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179052 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="nbdb" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179058 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="nbdb" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179064 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="sbdb" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179069 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="sbdb" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179080 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovn-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179086 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovn-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179117 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="northd" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179123 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="northd" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179129 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovn-acl-logging" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179136 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovn-acl-logging" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179247 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179279 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kube-rbac-proxy-node" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179287 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179293 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovn-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179301 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="northd" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179308 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="nbdb" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179315 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179322 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179352 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="sbdb" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179359 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovn-acl-logging" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179477 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179485 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.179516 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179523 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179634 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.179645 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerName="ovnkube-controller" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.181483 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183185 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-bin\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183215 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-slash\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183234 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183253 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-config\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183276 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-systemd-units\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183293 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-script-lib\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183311 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8490fb8-480e-4d81-82dd-4411b1f29278-ovn-node-metrics-cert\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183330 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-kubelet\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183347 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-env-overrides\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183367 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-openvswitch\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183380 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-netns\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183393 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-ovn-kubernetes\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183409 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-netd\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183432 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-var-lib-openvswitch\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183451 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-ovn\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183465 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-systemd\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183481 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-log-socket\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183499 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7hgf\" (UniqueName: \"kubernetes.io/projected/c8490fb8-480e-4d81-82dd-4411b1f29278-kube-api-access-r7hgf\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183517 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-node-log\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183530 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-etc-openvswitch\") pod \"c8490fb8-480e-4d81-82dd-4411b1f29278\" (UID: \"c8490fb8-480e-4d81-82dd-4411b1f29278\") " Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183601 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-run-netns\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183619 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovnkube-script-lib\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183636 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpzns\" (UniqueName: \"kubernetes.io/projected/a8f04d05-416b-4a8a-b8af-bb116810efd3-kube-api-access-wpzns\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183652 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-etc-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183666 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovnkube-config\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183681 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-systemd\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183706 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-cni-bin\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183723 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-var-lib-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183748 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-kubelet\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183769 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-cni-netd\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183789 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183804 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-env-overrides\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183824 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-slash\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183839 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-systemd-units\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183853 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-log-socket\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183885 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183900 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovn-node-metrics-cert\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183305 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-slash" (OuterVolumeSpecName: "host-slash") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183329 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183422 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183961 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183786 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183886 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183947 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-node-log" (OuterVolumeSpecName: "node-log") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183988 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184083 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184109 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184128 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184146 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184163 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184245 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184281 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.183916 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-ovn\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184369 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184410 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-node-log\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184456 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-log-socket" (OuterVolumeSpecName: "log-socket") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184496 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184503 4690 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184518 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184530 4690 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184540 4690 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184550 4690 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184558 4690 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8490fb8-480e-4d81-82dd-4411b1f29278-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184566 4690 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184577 4690 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184586 4690 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184594 4690 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184603 4690 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184613 4690 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184622 4690 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184630 4690 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.184638 4690 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.189977 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8490fb8-480e-4d81-82dd-4411b1f29278-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.190138 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8490fb8-480e-4d81-82dd-4411b1f29278-kube-api-access-r7hgf" (OuterVolumeSpecName: "kube-api-access-r7hgf") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "kube-api-access-r7hgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.198897 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c8490fb8-480e-4d81-82dd-4411b1f29278" (UID: "c8490fb8-480e-4d81-82dd-4411b1f29278"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.285533 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-node-log\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.285648 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-run-netns\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.285675 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovnkube-script-lib\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.285750 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpzns\" (UniqueName: \"kubernetes.io/projected/a8f04d05-416b-4a8a-b8af-bb116810efd3-kube-api-access-wpzns\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.285774 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-etc-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.286361 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovnkube-config\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.285765 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-node-log\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.286123 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-run-netns\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.286674 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovnkube-script-lib\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.285885 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-etc-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287153 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovnkube-config\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287205 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-systemd\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287227 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-cni-bin\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287246 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-var-lib-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287284 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-kubelet\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287315 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-cni-netd\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287344 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287364 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-env-overrides\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287390 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-slash\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287419 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-systemd-units\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287435 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-log-socket\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287453 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287472 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovn-node-metrics-cert\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287492 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-ovn\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287518 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287557 4690 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287570 4690 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287581 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7hgf\" (UniqueName: \"kubernetes.io/projected/c8490fb8-480e-4d81-82dd-4411b1f29278-kube-api-access-r7hgf\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287591 4690 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8490fb8-480e-4d81-82dd-4411b1f29278-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287601 4690 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8490fb8-480e-4d81-82dd-4411b1f29278-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287629 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287654 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-systemd\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287674 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-cni-bin\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287692 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-var-lib-openvswitch\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287713 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-kubelet\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287732 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-cni-netd\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287763 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287827 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-log-socket\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287888 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287905 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-host-slash\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.287930 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-run-ovn\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.288002 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a8f04d05-416b-4a8a-b8af-bb116810efd3-systemd-units\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.288358 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8f04d05-416b-4a8a-b8af-bb116810efd3-env-overrides\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.292519 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8f04d05-416b-4a8a-b8af-bb116810efd3-ovn-node-metrics-cert\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.302303 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpzns\" (UniqueName: \"kubernetes.io/projected/a8f04d05-416b-4a8a-b8af-bb116810efd3-kube-api-access-wpzns\") pod \"ovnkube-node-2h5pp\" (UID: \"a8f04d05-416b-4a8a-b8af-bb116810efd3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.316619 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovnkube-controller/3.log" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.319476 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovn-acl-logging/0.log" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320090 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vmddn_c8490fb8-480e-4d81-82dd-4411b1f29278/ovn-controller/0.log" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320626 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" exitCode=0 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320660 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" exitCode=0 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320668 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" exitCode=0 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320677 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3" exitCode=0 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320684 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555" exitCode=0 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320692 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a" exitCode=0 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320699 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628" exitCode=143 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320706 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8490fb8-480e-4d81-82dd-4411b1f29278" containerID="4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf" exitCode=143 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320721 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320801 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320832 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320843 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320853 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320886 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320897 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320908 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320918 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320924 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320929 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320935 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320941 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320947 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320952 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320957 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320966 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320975 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320981 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320987 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320992 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.320998 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321003 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321012 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321018 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321023 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321029 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321036 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321045 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321051 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321057 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321062 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321063 4690 scope.go:117] "RemoveContainer" containerID="950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321067 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321165 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321184 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321192 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321199 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321205 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321225 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vmddn" event={"ID":"c8490fb8-480e-4d81-82dd-4411b1f29278","Type":"ContainerDied","Data":"6c2a01d440258f2faa2abbb255797674c12b4118b69c7fc65a55cf7500b4df35"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321246 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321252 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321258 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321264 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321269 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321275 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321283 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321288 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321294 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.321302 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.323924 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/2.log" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.325379 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/1.log" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.325421 4690 generic.go:334] "Generic (PLEG): container finished" podID="bf580211-e03a-444e-98b8-c12f99091b80" containerID="ffc489323eede1d26eeba3894b5d1ebc6e278d6fb487a0a0ce651b5a1d2f456f" exitCode=2 Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.325461 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tf2cb" event={"ID":"bf580211-e03a-444e-98b8-c12f99091b80","Type":"ContainerDied","Data":"ffc489323eede1d26eeba3894b5d1ebc6e278d6fb487a0a0ce651b5a1d2f456f"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.325481 4690 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6"} Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.326495 4690 scope.go:117] "RemoveContainer" containerID="ffc489323eede1d26eeba3894b5d1ebc6e278d6fb487a0a0ce651b5a1d2f456f" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.326782 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-tf2cb_openshift-multus(bf580211-e03a-444e-98b8-c12f99091b80)\"" pod="openshift-multus/multus-tf2cb" podUID="bf580211-e03a-444e-98b8-c12f99091b80" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.345540 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.366181 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vmddn"] Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.369927 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vmddn"] Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.373136 4690 scope.go:117] "RemoveContainer" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.386560 4690 scope.go:117] "RemoveContainer" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.398416 4690 scope.go:117] "RemoveContainer" containerID="79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.409794 4690 scope.go:117] "RemoveContainer" containerID="c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.423272 4690 scope.go:117] "RemoveContainer" containerID="53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.467725 4690 scope.go:117] "RemoveContainer" containerID="fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.481368 4690 scope.go:117] "RemoveContainer" containerID="4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.498046 4690 scope.go:117] "RemoveContainer" containerID="4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.504239 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.515049 4690 scope.go:117] "RemoveContainer" containerID="950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.515419 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": container with ID starting with 950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79 not found: ID does not exist" containerID="950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.515455 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} err="failed to get container status \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": rpc error: code = NotFound desc = could not find container \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": container with ID starting with 950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.515479 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.515823 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": container with ID starting with dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc not found: ID does not exist" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.515850 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} err="failed to get container status \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": rpc error: code = NotFound desc = could not find container \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": container with ID starting with dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.515876 4690 scope.go:117] "RemoveContainer" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.516213 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": container with ID starting with 8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8 not found: ID does not exist" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.516252 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} err="failed to get container status \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": rpc error: code = NotFound desc = could not find container \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": container with ID starting with 8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.516274 4690 scope.go:117] "RemoveContainer" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.516526 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": container with ID starting with 149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef not found: ID does not exist" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.516556 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} err="failed to get container status \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": rpc error: code = NotFound desc = could not find container \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": container with ID starting with 149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.516574 4690 scope.go:117] "RemoveContainer" containerID="79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.516775 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": container with ID starting with 79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3 not found: ID does not exist" containerID="79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.516798 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} err="failed to get container status \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": rpc error: code = NotFound desc = could not find container \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": container with ID starting with 79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.516814 4690 scope.go:117] "RemoveContainer" containerID="c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.517037 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": container with ID starting with c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555 not found: ID does not exist" containerID="c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.517108 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} err="failed to get container status \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": rpc error: code = NotFound desc = could not find container \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": container with ID starting with c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.517143 4690 scope.go:117] "RemoveContainer" containerID="53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.517387 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": container with ID starting with 53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a not found: ID does not exist" containerID="53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.517408 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} err="failed to get container status \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": rpc error: code = NotFound desc = could not find container \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": container with ID starting with 53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.517421 4690 scope.go:117] "RemoveContainer" containerID="fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.517620 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": container with ID starting with fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628 not found: ID does not exist" containerID="fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.517643 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} err="failed to get container status \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": rpc error: code = NotFound desc = could not find container \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": container with ID starting with fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.517662 4690 scope.go:117] "RemoveContainer" containerID="4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.518040 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": container with ID starting with 4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf not found: ID does not exist" containerID="4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.518070 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} err="failed to get container status \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": rpc error: code = NotFound desc = could not find container \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": container with ID starting with 4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.518089 4690 scope.go:117] "RemoveContainer" containerID="4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647" Dec 03 13:21:05 crc kubenswrapper[4690]: E1203 13:21:05.518423 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": container with ID starting with 4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647 not found: ID does not exist" containerID="4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.518470 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} err="failed to get container status \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": rpc error: code = NotFound desc = could not find container \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": container with ID starting with 4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.518500 4690 scope.go:117] "RemoveContainer" containerID="950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.518762 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} err="failed to get container status \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": rpc error: code = NotFound desc = could not find container \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": container with ID starting with 950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.518791 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.519027 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} err="failed to get container status \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": rpc error: code = NotFound desc = could not find container \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": container with ID starting with dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.519067 4690 scope.go:117] "RemoveContainer" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.519344 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} err="failed to get container status \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": rpc error: code = NotFound desc = could not find container \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": container with ID starting with 8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.519372 4690 scope.go:117] "RemoveContainer" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.519718 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} err="failed to get container status \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": rpc error: code = NotFound desc = could not find container \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": container with ID starting with 149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.519957 4690 scope.go:117] "RemoveContainer" containerID="79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.520491 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} err="failed to get container status \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": rpc error: code = NotFound desc = could not find container \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": container with ID starting with 79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.520523 4690 scope.go:117] "RemoveContainer" containerID="c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.522108 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} err="failed to get container status \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": rpc error: code = NotFound desc = could not find container \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": container with ID starting with c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.522167 4690 scope.go:117] "RemoveContainer" containerID="53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.522550 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} err="failed to get container status \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": rpc error: code = NotFound desc = could not find container \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": container with ID starting with 53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.522580 4690 scope.go:117] "RemoveContainer" containerID="fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.523211 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} err="failed to get container status \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": rpc error: code = NotFound desc = could not find container \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": container with ID starting with fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.523320 4690 scope.go:117] "RemoveContainer" containerID="4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.524365 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} err="failed to get container status \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": rpc error: code = NotFound desc = could not find container \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": container with ID starting with 4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.524394 4690 scope.go:117] "RemoveContainer" containerID="4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.524794 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} err="failed to get container status \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": rpc error: code = NotFound desc = could not find container \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": container with ID starting with 4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.524827 4690 scope.go:117] "RemoveContainer" containerID="950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.525440 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} err="failed to get container status \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": rpc error: code = NotFound desc = could not find container \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": container with ID starting with 950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.525479 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.525956 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} err="failed to get container status \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": rpc error: code = NotFound desc = could not find container \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": container with ID starting with dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.525989 4690 scope.go:117] "RemoveContainer" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.526471 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} err="failed to get container status \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": rpc error: code = NotFound desc = could not find container \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": container with ID starting with 8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.526528 4690 scope.go:117] "RemoveContainer" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.527207 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} err="failed to get container status \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": rpc error: code = NotFound desc = could not find container \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": container with ID starting with 149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.527246 4690 scope.go:117] "RemoveContainer" containerID="79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.527649 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} err="failed to get container status \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": rpc error: code = NotFound desc = could not find container \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": container with ID starting with 79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.527681 4690 scope.go:117] "RemoveContainer" containerID="c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.528114 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} err="failed to get container status \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": rpc error: code = NotFound desc = could not find container \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": container with ID starting with c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.528142 4690 scope.go:117] "RemoveContainer" containerID="53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.528625 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} err="failed to get container status \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": rpc error: code = NotFound desc = could not find container \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": container with ID starting with 53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.528660 4690 scope.go:117] "RemoveContainer" containerID="fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.529035 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} err="failed to get container status \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": rpc error: code = NotFound desc = could not find container \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": container with ID starting with fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.529067 4690 scope.go:117] "RemoveContainer" containerID="4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.529486 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} err="failed to get container status \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": rpc error: code = NotFound desc = could not find container \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": container with ID starting with 4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.529513 4690 scope.go:117] "RemoveContainer" containerID="4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.529844 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} err="failed to get container status \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": rpc error: code = NotFound desc = could not find container \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": container with ID starting with 4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.529864 4690 scope.go:117] "RemoveContainer" containerID="950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.530375 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} err="failed to get container status \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": rpc error: code = NotFound desc = could not find container \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": container with ID starting with 950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.530396 4690 scope.go:117] "RemoveContainer" containerID="dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.530824 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc"} err="failed to get container status \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": rpc error: code = NotFound desc = could not find container \"dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc\": container with ID starting with dd36893249b8927e2cc8a575a67c6b39552a5e7c3d3e3977c949dbd212a9ddcc not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.530908 4690 scope.go:117] "RemoveContainer" containerID="8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.531350 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8"} err="failed to get container status \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": rpc error: code = NotFound desc = could not find container \"8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8\": container with ID starting with 8b9bce4e5080b68ddf8895d476b6c5bdaa4519a0b160aba175dccc295a7384a8 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.531396 4690 scope.go:117] "RemoveContainer" containerID="149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.531749 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef"} err="failed to get container status \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": rpc error: code = NotFound desc = could not find container \"149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef\": container with ID starting with 149bcc570d919060865462739d0d736a25e048ee32e9d10ce28b3ab848f5b1ef not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.531774 4690 scope.go:117] "RemoveContainer" containerID="79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.532124 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3"} err="failed to get container status \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": rpc error: code = NotFound desc = could not find container \"79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3\": container with ID starting with 79fe27dfb449be7ec497051bde5b31757c9517d10ef5dc220d735adf16eaffe3 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.532156 4690 scope.go:117] "RemoveContainer" containerID="c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.532500 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555"} err="failed to get container status \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": rpc error: code = NotFound desc = could not find container \"c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555\": container with ID starting with c71ed74a4242173cc41c6a67840db4ba486e796d6385d599b7bd918af7470555 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.532543 4690 scope.go:117] "RemoveContainer" containerID="53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.533017 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a"} err="failed to get container status \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": rpc error: code = NotFound desc = could not find container \"53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a\": container with ID starting with 53db13b0d8c311b01de7dea89f7cfdca017d5cfb07ece9be148e4f695334a23a not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.533045 4690 scope.go:117] "RemoveContainer" containerID="fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.533377 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628"} err="failed to get container status \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": rpc error: code = NotFound desc = could not find container \"fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628\": container with ID starting with fd35dc7bdf9390fcf38013299fc76dd1b2d26bb75f0e3b267f768814a041f628 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.533428 4690 scope.go:117] "RemoveContainer" containerID="4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.534285 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf"} err="failed to get container status \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": rpc error: code = NotFound desc = could not find container \"4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf\": container with ID starting with 4afa50d60783689440b5a18dcd04150ec9ca1fdba72ffa7e1b9b325bebe54aaf not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.534319 4690 scope.go:117] "RemoveContainer" containerID="4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.534838 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647"} err="failed to get container status \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": rpc error: code = NotFound desc = could not find container \"4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647\": container with ID starting with 4fd7d7396c62bb7fad936ca006dae7e111e0ece0371b2c539f699c33d10ee647 not found: ID does not exist" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.534899 4690 scope.go:117] "RemoveContainer" containerID="950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79" Dec 03 13:21:05 crc kubenswrapper[4690]: I1203 13:21:05.535557 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79"} err="failed to get container status \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": rpc error: code = NotFound desc = could not find container \"950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79\": container with ID starting with 950b11498e29b10982f6695ffc8c7db6dfd957cd85100b7a347e89ddad019b79 not found: ID does not exist" Dec 03 13:21:06 crc kubenswrapper[4690]: I1203 13:21:06.328174 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8490fb8-480e-4d81-82dd-4411b1f29278" path="/var/lib/kubelet/pods/c8490fb8-480e-4d81-82dd-4411b1f29278/volumes" Dec 03 13:21:06 crc kubenswrapper[4690]: I1203 13:21:06.331438 4690 generic.go:334] "Generic (PLEG): container finished" podID="a8f04d05-416b-4a8a-b8af-bb116810efd3" containerID="f4b0296b6dd1146a74d9cf1761ba4f36ecea530efced40ec803df433b4837986" exitCode=0 Dec 03 13:21:06 crc kubenswrapper[4690]: I1203 13:21:06.331483 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerDied","Data":"f4b0296b6dd1146a74d9cf1761ba4f36ecea530efced40ec803df433b4837986"} Dec 03 13:21:06 crc kubenswrapper[4690]: I1203 13:21:06.331530 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"51178b78dc92ba5eeff0334f07f964ad2b0b1dc907a1408a57da40bf094008eb"} Dec 03 13:21:07 crc kubenswrapper[4690]: I1203 13:21:07.344169 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"327b8aec3d7b310e931b69cfb93f63d2ac1379509a5602c2087fe7eea65815e4"} Dec 03 13:21:07 crc kubenswrapper[4690]: I1203 13:21:07.344537 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"1805ed8aed7bf05b61cc2b3df422897e4b2c6b8b8692dc42eb8e977c54324e33"} Dec 03 13:21:07 crc kubenswrapper[4690]: I1203 13:21:07.344558 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"8a11cfbe2d653469ef024c52d3da778853d028159d8c1d6b48502c7a2cc4d0dc"} Dec 03 13:21:07 crc kubenswrapper[4690]: I1203 13:21:07.344571 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"5ea70a850bc28efdd9f907ee6992ccaaa1c0a42ebf14cb4f1233fb3d635192e0"} Dec 03 13:21:07 crc kubenswrapper[4690]: I1203 13:21:07.344585 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"66144898db520b39c1c9cf112e9ed82af569f2c5f07c5c6f0b02b4cdf9956e16"} Dec 03 13:21:07 crc kubenswrapper[4690]: I1203 13:21:07.344597 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"45dda0f330055e9183fe1319471d6b9d52c729721ce21611eb8fcd2f666f868f"} Dec 03 13:21:09 crc kubenswrapper[4690]: I1203 13:21:09.361496 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"edae48ca3dfb24f6bb33affed9acfe056280997ff63b82ff2b97d0085531dfae"} Dec 03 13:21:12 crc kubenswrapper[4690]: I1203 13:21:12.384812 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" event={"ID":"a8f04d05-416b-4a8a-b8af-bb116810efd3","Type":"ContainerStarted","Data":"984644da8aabcd3342da001a6dcf46b00fb7c41e32ad00c6a6efa4831e85460e"} Dec 03 13:21:12 crc kubenswrapper[4690]: I1203 13:21:12.385613 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:12 crc kubenswrapper[4690]: I1203 13:21:12.385630 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:12 crc kubenswrapper[4690]: I1203 13:21:12.385642 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:12 crc kubenswrapper[4690]: I1203 13:21:12.411497 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:12 crc kubenswrapper[4690]: I1203 13:21:12.416543 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:12 crc kubenswrapper[4690]: I1203 13:21:12.418362 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" podStartSLOduration=7.418341863 podStartE2EDuration="7.418341863s" podCreationTimestamp="2025-12-03 13:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:21:12.417032069 +0000 UTC m=+718.397952532" watchObservedRunningTime="2025-12-03 13:21:12.418341863 +0000 UTC m=+718.399262296" Dec 03 13:21:14 crc kubenswrapper[4690]: I1203 13:21:14.556180 4690 scope.go:117] "RemoveContainer" containerID="f0107e26fcb557ce4cb52f17cb3a82f53766fe938bb3ce2d994b35de5cb009c6" Dec 03 13:21:14 crc kubenswrapper[4690]: I1203 13:21:14.865278 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx"] Dec 03 13:21:14 crc kubenswrapper[4690]: I1203 13:21:14.866753 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:14 crc kubenswrapper[4690]: I1203 13:21:14.873322 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 13:21:14 crc kubenswrapper[4690]: I1203 13:21:14.874254 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx"] Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.022799 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.022891 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.022926 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzghp\" (UniqueName: \"kubernetes.io/projected/2d467dd5-228e-436a-8fa0-c96bbf0ce110-kube-api-access-kzghp\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.124403 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.124502 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.124559 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzghp\" (UniqueName: \"kubernetes.io/projected/2d467dd5-228e-436a-8fa0-c96bbf0ce110-kube-api-access-kzghp\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.125288 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.125463 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.156454 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzghp\" (UniqueName: \"kubernetes.io/projected/2d467dd5-228e-436a-8fa0-c96bbf0ce110-kube-api-access-kzghp\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.188231 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: E1203 13:21:15.219123 4690 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace_2d467dd5-228e-436a-8fa0-c96bbf0ce110_0(4080aefa818645d85b60bd0c53e699f8bd29d102b3c872ab4fd8059aba3d3cc8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 13:21:15 crc kubenswrapper[4690]: E1203 13:21:15.219396 4690 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace_2d467dd5-228e-436a-8fa0-c96bbf0ce110_0(4080aefa818645d85b60bd0c53e699f8bd29d102b3c872ab4fd8059aba3d3cc8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: E1203 13:21:15.219533 4690 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace_2d467dd5-228e-436a-8fa0-c96bbf0ce110_0(4080aefa818645d85b60bd0c53e699f8bd29d102b3c872ab4fd8059aba3d3cc8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: E1203 13:21:15.219719 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace(2d467dd5-228e-436a-8fa0-c96bbf0ce110)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace(2d467dd5-228e-436a-8fa0-c96bbf0ce110)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace_2d467dd5-228e-436a-8fa0-c96bbf0ce110_0(4080aefa818645d85b60bd0c53e699f8bd29d102b3c872ab4fd8059aba3d3cc8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.407855 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/2.log" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.408080 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: I1203 13:21:15.408817 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: E1203 13:21:15.440494 4690 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace_2d467dd5-228e-436a-8fa0-c96bbf0ce110_0(53870483e53914dbc1716b285270942478541346ac967e5bdcf0f2cae91dffc0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 13:21:15 crc kubenswrapper[4690]: E1203 13:21:15.440624 4690 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace_2d467dd5-228e-436a-8fa0-c96bbf0ce110_0(53870483e53914dbc1716b285270942478541346ac967e5bdcf0f2cae91dffc0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: E1203 13:21:15.440662 4690 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace_2d467dd5-228e-436a-8fa0-c96bbf0ce110_0(53870483e53914dbc1716b285270942478541346ac967e5bdcf0f2cae91dffc0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:15 crc kubenswrapper[4690]: E1203 13:21:15.440744 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace(2d467dd5-228e-436a-8fa0-c96bbf0ce110)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace(2d467dd5-228e-436a-8fa0-c96bbf0ce110)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_openshift-marketplace_2d467dd5-228e-436a-8fa0-c96bbf0ce110_0(53870483e53914dbc1716b285270942478541346ac967e5bdcf0f2cae91dffc0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" Dec 03 13:21:19 crc kubenswrapper[4690]: I1203 13:21:19.315795 4690 scope.go:117] "RemoveContainer" containerID="ffc489323eede1d26eeba3894b5d1ebc6e278d6fb487a0a0ce651b5a1d2f456f" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.268104 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx"] Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.270314 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.277568 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx"] Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.396711 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hwvb\" (UniqueName: \"kubernetes.io/projected/d67d533a-906d-49f6-8331-7bc5f85bff55-kube-api-access-8hwvb\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.397331 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.397456 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.438858 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tf2cb_bf580211-e03a-444e-98b8-c12f99091b80/kube-multus/2.log" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.439028 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tf2cb" event={"ID":"bf580211-e03a-444e-98b8-c12f99091b80","Type":"ContainerStarted","Data":"e269d8df78e5c2e0afffbeabd3b3f58914967c2e2d852ac23e23ae91e9b15f9b"} Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.498428 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hwvb\" (UniqueName: \"kubernetes.io/projected/d67d533a-906d-49f6-8331-7bc5f85bff55-kube-api-access-8hwvb\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.498525 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.498648 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.499498 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.499562 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.538361 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hwvb\" (UniqueName: \"kubernetes.io/projected/d67d533a-906d-49f6-8331-7bc5f85bff55-kube-api-access-8hwvb\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: I1203 13:21:20.597951 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: E1203 13:21:20.630982 4690 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_openshift-marketplace_d67d533a-906d-49f6-8331-7bc5f85bff55_0(a1be267f3efa8ab1420238ac6236f6f602b20c8565d610c26707a71dd4a69b8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 13:21:20 crc kubenswrapper[4690]: E1203 13:21:20.631077 4690 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_openshift-marketplace_d67d533a-906d-49f6-8331-7bc5f85bff55_0(a1be267f3efa8ab1420238ac6236f6f602b20c8565d610c26707a71dd4a69b8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: E1203 13:21:20.631104 4690 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_openshift-marketplace_d67d533a-906d-49f6-8331-7bc5f85bff55_0(a1be267f3efa8ab1420238ac6236f6f602b20c8565d610c26707a71dd4a69b8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:20 crc kubenswrapper[4690]: E1203 13:21:20.631161 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_openshift-marketplace(d67d533a-906d-49f6-8331-7bc5f85bff55)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_openshift-marketplace(d67d533a-906d-49f6-8331-7bc5f85bff55)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_openshift-marketplace_d67d533a-906d-49f6-8331-7bc5f85bff55_0(a1be267f3efa8ab1420238ac6236f6f602b20c8565d610c26707a71dd4a69b8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" podUID="d67d533a-906d-49f6-8331-7bc5f85bff55" Dec 03 13:21:21 crc kubenswrapper[4690]: I1203 13:21:21.443237 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:21 crc kubenswrapper[4690]: I1203 13:21:21.443711 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:21 crc kubenswrapper[4690]: I1203 13:21:21.624168 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx"] Dec 03 13:21:21 crc kubenswrapper[4690]: W1203 13:21:21.629596 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd67d533a_906d_49f6_8331_7bc5f85bff55.slice/crio-69c57ae41e350ca37d2ae2ab69f2fa6526852375f0997f82771bfbbe50456d6b WatchSource:0}: Error finding container 69c57ae41e350ca37d2ae2ab69f2fa6526852375f0997f82771bfbbe50456d6b: Status 404 returned error can't find the container with id 69c57ae41e350ca37d2ae2ab69f2fa6526852375f0997f82771bfbbe50456d6b Dec 03 13:21:22 crc kubenswrapper[4690]: I1203 13:21:22.450216 4690 generic.go:334] "Generic (PLEG): container finished" podID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerID="e19c615edf6323751868af1b57e25b77a0966b7207d4b653565720dd1dfe2acb" exitCode=0 Dec 03 13:21:22 crc kubenswrapper[4690]: I1203 13:21:22.450300 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" event={"ID":"d67d533a-906d-49f6-8331-7bc5f85bff55","Type":"ContainerDied","Data":"e19c615edf6323751868af1b57e25b77a0966b7207d4b653565720dd1dfe2acb"} Dec 03 13:21:22 crc kubenswrapper[4690]: I1203 13:21:22.450540 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" event={"ID":"d67d533a-906d-49f6-8331-7bc5f85bff55","Type":"ContainerStarted","Data":"69c57ae41e350ca37d2ae2ab69f2fa6526852375f0997f82771bfbbe50456d6b"} Dec 03 13:21:24 crc kubenswrapper[4690]: I1203 13:21:24.462834 4690 generic.go:334] "Generic (PLEG): container finished" podID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerID="52e4f4bc7a1311a2b5997fdb2ff735563885eccec27fba5e91d06f1d3c905497" exitCode=0 Dec 03 13:21:24 crc kubenswrapper[4690]: I1203 13:21:24.462969 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" event={"ID":"d67d533a-906d-49f6-8331-7bc5f85bff55","Type":"ContainerDied","Data":"52e4f4bc7a1311a2b5997fdb2ff735563885eccec27fba5e91d06f1d3c905497"} Dec 03 13:21:25 crc kubenswrapper[4690]: I1203 13:21:25.474904 4690 generic.go:334] "Generic (PLEG): container finished" podID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerID="8e5aec5ce882c386380a1c2cb1d615fd964ee27e5e18e11386a564c601d22c2b" exitCode=0 Dec 03 13:21:25 crc kubenswrapper[4690]: I1203 13:21:25.474984 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" event={"ID":"d67d533a-906d-49f6-8331-7bc5f85bff55","Type":"ContainerDied","Data":"8e5aec5ce882c386380a1c2cb1d615fd964ee27e5e18e11386a564c601d22c2b"} Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.718105 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.792136 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-util\") pod \"d67d533a-906d-49f6-8331-7bc5f85bff55\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.792200 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-bundle\") pod \"d67d533a-906d-49f6-8331-7bc5f85bff55\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.792257 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hwvb\" (UniqueName: \"kubernetes.io/projected/d67d533a-906d-49f6-8331-7bc5f85bff55-kube-api-access-8hwvb\") pod \"d67d533a-906d-49f6-8331-7bc5f85bff55\" (UID: \"d67d533a-906d-49f6-8331-7bc5f85bff55\") " Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.793193 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-bundle" (OuterVolumeSpecName: "bundle") pod "d67d533a-906d-49f6-8331-7bc5f85bff55" (UID: "d67d533a-906d-49f6-8331-7bc5f85bff55"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.798071 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d67d533a-906d-49f6-8331-7bc5f85bff55-kube-api-access-8hwvb" (OuterVolumeSpecName: "kube-api-access-8hwvb") pod "d67d533a-906d-49f6-8331-7bc5f85bff55" (UID: "d67d533a-906d-49f6-8331-7bc5f85bff55"). InnerVolumeSpecName "kube-api-access-8hwvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.805262 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-util" (OuterVolumeSpecName: "util") pod "d67d533a-906d-49f6-8331-7bc5f85bff55" (UID: "d67d533a-906d-49f6-8331-7bc5f85bff55"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.893739 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-util\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.893773 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d67d533a-906d-49f6-8331-7bc5f85bff55-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:26 crc kubenswrapper[4690]: I1203 13:21:26.893784 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hwvb\" (UniqueName: \"kubernetes.io/projected/d67d533a-906d-49f6-8331-7bc5f85bff55-kube-api-access-8hwvb\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:27 crc kubenswrapper[4690]: I1203 13:21:27.493963 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" event={"ID":"d67d533a-906d-49f6-8331-7bc5f85bff55","Type":"ContainerDied","Data":"69c57ae41e350ca37d2ae2ab69f2fa6526852375f0997f82771bfbbe50456d6b"} Dec 03 13:21:27 crc kubenswrapper[4690]: I1203 13:21:27.494025 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69c57ae41e350ca37d2ae2ab69f2fa6526852375f0997f82771bfbbe50456d6b" Dec 03 13:21:27 crc kubenswrapper[4690]: I1203 13:21:27.494049 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx" Dec 03 13:21:30 crc kubenswrapper[4690]: I1203 13:21:30.314747 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:30 crc kubenswrapper[4690]: I1203 13:21:30.315710 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:30 crc kubenswrapper[4690]: I1203 13:21:30.564376 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx"] Dec 03 13:21:30 crc kubenswrapper[4690]: W1203 13:21:30.570909 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d467dd5_228e_436a_8fa0_c96bbf0ce110.slice/crio-bb5d775950740a00f62c91f87bff7e06ed327a9b68cd67d12cc1873bfd13e751 WatchSource:0}: Error finding container bb5d775950740a00f62c91f87bff7e06ed327a9b68cd67d12cc1873bfd13e751: Status 404 returned error can't find the container with id bb5d775950740a00f62c91f87bff7e06ed327a9b68cd67d12cc1873bfd13e751 Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.518107 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" event={"ID":"2d467dd5-228e-436a-8fa0-c96bbf0ce110","Type":"ContainerStarted","Data":"bb5d775950740a00f62c91f87bff7e06ed327a9b68cd67d12cc1873bfd13e751"} Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.659520 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lqf4z"] Dec 03 13:21:31 crc kubenswrapper[4690]: E1203 13:21:31.659743 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerName="extract" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.659755 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerName="extract" Dec 03 13:21:31 crc kubenswrapper[4690]: E1203 13:21:31.659772 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerName="util" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.659779 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerName="util" Dec 03 13:21:31 crc kubenswrapper[4690]: E1203 13:21:31.659787 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerName="pull" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.659794 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerName="pull" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.659962 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67d533a-906d-49f6-8331-7bc5f85bff55" containerName="extract" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.660333 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-lqf4z" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.662817 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.663169 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.663944 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-7pzxg" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.684051 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lqf4z"] Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.769712 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdw9q\" (UniqueName: \"kubernetes.io/projected/fb39061a-2c4c-4c29-95b6-6f070367df86-kube-api-access-mdw9q\") pod \"cluster-logging-operator-ff9846bd-lqf4z\" (UID: \"fb39061a-2c4c-4c29-95b6-6f070367df86\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lqf4z" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.871008 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdw9q\" (UniqueName: \"kubernetes.io/projected/fb39061a-2c4c-4c29-95b6-6f070367df86-kube-api-access-mdw9q\") pod \"cluster-logging-operator-ff9846bd-lqf4z\" (UID: \"fb39061a-2c4c-4c29-95b6-6f070367df86\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lqf4z" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.899266 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdw9q\" (UniqueName: \"kubernetes.io/projected/fb39061a-2c4c-4c29-95b6-6f070367df86-kube-api-access-mdw9q\") pod \"cluster-logging-operator-ff9846bd-lqf4z\" (UID: \"fb39061a-2c4c-4c29-95b6-6f070367df86\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lqf4z" Dec 03 13:21:31 crc kubenswrapper[4690]: I1203 13:21:31.976142 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-lqf4z" Dec 03 13:21:32 crc kubenswrapper[4690]: I1203 13:21:32.168235 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lqf4z"] Dec 03 13:21:32 crc kubenswrapper[4690]: W1203 13:21:32.179268 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb39061a_2c4c_4c29_95b6_6f070367df86.slice/crio-00abbc8d230b4dcf16f314ac3bb2c6befb9fa9ef775aa717763b84cf562df626 WatchSource:0}: Error finding container 00abbc8d230b4dcf16f314ac3bb2c6befb9fa9ef775aa717763b84cf562df626: Status 404 returned error can't find the container with id 00abbc8d230b4dcf16f314ac3bb2c6befb9fa9ef775aa717763b84cf562df626 Dec 03 13:21:32 crc kubenswrapper[4690]: I1203 13:21:32.527667 4690 generic.go:334] "Generic (PLEG): container finished" podID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerID="d68bd61a91d4c621a8d9dfae4d4d4d28380fc75ca2c2a16a0552486032db7762" exitCode=0 Dec 03 13:21:32 crc kubenswrapper[4690]: I1203 13:21:32.527728 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" event={"ID":"2d467dd5-228e-436a-8fa0-c96bbf0ce110","Type":"ContainerDied","Data":"d68bd61a91d4c621a8d9dfae4d4d4d28380fc75ca2c2a16a0552486032db7762"} Dec 03 13:21:32 crc kubenswrapper[4690]: I1203 13:21:32.529362 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-lqf4z" event={"ID":"fb39061a-2c4c-4c29-95b6-6f070367df86","Type":"ContainerStarted","Data":"00abbc8d230b4dcf16f314ac3bb2c6befb9fa9ef775aa717763b84cf562df626"} Dec 03 13:21:34 crc kubenswrapper[4690]: I1203 13:21:34.544295 4690 generic.go:334] "Generic (PLEG): container finished" podID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerID="05bea3f06c1a8d295696f7dfa31e480d530b4802d0bb451eaf51058fc9a33814" exitCode=0 Dec 03 13:21:34 crc kubenswrapper[4690]: I1203 13:21:34.544376 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" event={"ID":"2d467dd5-228e-436a-8fa0-c96bbf0ce110","Type":"ContainerDied","Data":"05bea3f06c1a8d295696f7dfa31e480d530b4802d0bb451eaf51058fc9a33814"} Dec 03 13:21:35 crc kubenswrapper[4690]: I1203 13:21:35.530006 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2h5pp" Dec 03 13:21:35 crc kubenswrapper[4690]: I1203 13:21:35.553238 4690 generic.go:334] "Generic (PLEG): container finished" podID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerID="884563a2ac50bd378133b0b9a38d794b239ca69ebcd76b40ed98d08131e7c3b1" exitCode=0 Dec 03 13:21:35 crc kubenswrapper[4690]: I1203 13:21:35.553296 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" event={"ID":"2d467dd5-228e-436a-8fa0-c96bbf0ce110","Type":"ContainerDied","Data":"884563a2ac50bd378133b0b9a38d794b239ca69ebcd76b40ed98d08131e7c3b1"} Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.539592 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.570170 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzghp\" (UniqueName: \"kubernetes.io/projected/2d467dd5-228e-436a-8fa0-c96bbf0ce110-kube-api-access-kzghp\") pod \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.570266 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-util\") pod \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.570320 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-bundle\") pod \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\" (UID: \"2d467dd5-228e-436a-8fa0-c96bbf0ce110\") " Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.571844 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-bundle" (OuterVolumeSpecName: "bundle") pod "2d467dd5-228e-436a-8fa0-c96bbf0ce110" (UID: "2d467dd5-228e-436a-8fa0-c96bbf0ce110"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.584821 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-util" (OuterVolumeSpecName: "util") pod "2d467dd5-228e-436a-8fa0-c96bbf0ce110" (UID: "2d467dd5-228e-436a-8fa0-c96bbf0ce110"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.585303 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" event={"ID":"2d467dd5-228e-436a-8fa0-c96bbf0ce110","Type":"ContainerDied","Data":"bb5d775950740a00f62c91f87bff7e06ed327a9b68cd67d12cc1873bfd13e751"} Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.585361 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb5d775950740a00f62c91f87bff7e06ed327a9b68cd67d12cc1873bfd13e751" Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.585408 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx" Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.589843 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d467dd5-228e-436a-8fa0-c96bbf0ce110-kube-api-access-kzghp" (OuterVolumeSpecName: "kube-api-access-kzghp") pod "2d467dd5-228e-436a-8fa0-c96bbf0ce110" (UID: "2d467dd5-228e-436a-8fa0-c96bbf0ce110"). InnerVolumeSpecName "kube-api-access-kzghp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.672445 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzghp\" (UniqueName: \"kubernetes.io/projected/2d467dd5-228e-436a-8fa0-c96bbf0ce110-kube-api-access-kzghp\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.672480 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-util\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:38 crc kubenswrapper[4690]: I1203 13:21:38.672490 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d467dd5-228e-436a-8fa0-c96bbf0ce110-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:21:39 crc kubenswrapper[4690]: I1203 13:21:39.594862 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-lqf4z" event={"ID":"fb39061a-2c4c-4c29-95b6-6f070367df86","Type":"ContainerStarted","Data":"66dc7088a6224010def4ae07cc16ba17a3564d985345dded5c148f39c523adbe"} Dec 03 13:21:39 crc kubenswrapper[4690]: I1203 13:21:39.610486 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-lqf4z" podStartSLOduration=2.110121383 podStartE2EDuration="8.610467864s" podCreationTimestamp="2025-12-03 13:21:31 +0000 UTC" firstStartedPulling="2025-12-03 13:21:32.182042094 +0000 UTC m=+738.162962527" lastFinishedPulling="2025-12-03 13:21:38.682388575 +0000 UTC m=+744.663309008" observedRunningTime="2025-12-03 13:21:39.609410597 +0000 UTC m=+745.590331050" watchObservedRunningTime="2025-12-03 13:21:39.610467864 +0000 UTC m=+745.591388297" Dec 03 13:21:43 crc kubenswrapper[4690]: I1203 13:21:43.221294 4690 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.608948 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656"] Dec 03 13:21:50 crc kubenswrapper[4690]: E1203 13:21:50.610062 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerName="util" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.610078 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerName="util" Dec 03 13:21:50 crc kubenswrapper[4690]: E1203 13:21:50.610099 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerName="extract" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.610106 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerName="extract" Dec 03 13:21:50 crc kubenswrapper[4690]: E1203 13:21:50.610122 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerName="pull" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.610128 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerName="pull" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.610224 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d467dd5-228e-436a-8fa0-c96bbf0ce110" containerName="extract" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.610788 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.614549 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.615011 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.615324 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.616147 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.616588 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.624448 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-jzvlx" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.636789 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-webhook-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.636903 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-apiservice-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.636930 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vpn2\" (UniqueName: \"kubernetes.io/projected/72a896a4-41ed-4652-bfa5-358369eddfa2-kube-api-access-4vpn2\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.636981 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.637134 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/72a896a4-41ed-4652-bfa5-358369eddfa2-manager-config\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.653367 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656"] Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.738829 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-webhook-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.738935 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-apiservice-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.738969 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vpn2\" (UniqueName: \"kubernetes.io/projected/72a896a4-41ed-4652-bfa5-358369eddfa2-kube-api-access-4vpn2\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.739031 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.739058 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/72a896a4-41ed-4652-bfa5-358369eddfa2-manager-config\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.740154 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/72a896a4-41ed-4652-bfa5-358369eddfa2-manager-config\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.746852 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-webhook-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.747402 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.754411 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72a896a4-41ed-4652-bfa5-358369eddfa2-apiservice-cert\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.755611 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vpn2\" (UniqueName: \"kubernetes.io/projected/72a896a4-41ed-4652-bfa5-358369eddfa2-kube-api-access-4vpn2\") pod \"loki-operator-controller-manager-bb5b4b448-ql656\" (UID: \"72a896a4-41ed-4652-bfa5-358369eddfa2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:50 crc kubenswrapper[4690]: I1203 13:21:50.928006 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:21:51 crc kubenswrapper[4690]: I1203 13:21:51.202792 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656"] Dec 03 13:21:51 crc kubenswrapper[4690]: W1203 13:21:51.208416 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72a896a4_41ed_4652_bfa5_358369eddfa2.slice/crio-c9106b7adb1d8bdc1e738f498c1107cb40922a325dca45500adf41e333dd8157 WatchSource:0}: Error finding container c9106b7adb1d8bdc1e738f498c1107cb40922a325dca45500adf41e333dd8157: Status 404 returned error can't find the container with id c9106b7adb1d8bdc1e738f498c1107cb40922a325dca45500adf41e333dd8157 Dec 03 13:21:51 crc kubenswrapper[4690]: I1203 13:21:51.660836 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" event={"ID":"72a896a4-41ed-4652-bfa5-358369eddfa2","Type":"ContainerStarted","Data":"c9106b7adb1d8bdc1e738f498c1107cb40922a325dca45500adf41e333dd8157"} Dec 03 13:21:55 crc kubenswrapper[4690]: I1203 13:21:55.685088 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" event={"ID":"72a896a4-41ed-4652-bfa5-358369eddfa2","Type":"ContainerStarted","Data":"d5bac0e1a4b92e09c682d6a6c73639d04b653ebd02ba40a1b9d356d6fedb897a"} Dec 03 13:22:02 crc kubenswrapper[4690]: I1203 13:22:02.730278 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" event={"ID":"72a896a4-41ed-4652-bfa5-358369eddfa2","Type":"ContainerStarted","Data":"61e372a39daedf7c10891d82ef51aba626f4c614ffa67b80385e6da398cec57a"} Dec 03 13:22:02 crc kubenswrapper[4690]: I1203 13:22:02.730807 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:22:02 crc kubenswrapper[4690]: I1203 13:22:02.732311 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" Dec 03 13:22:02 crc kubenswrapper[4690]: I1203 13:22:02.760781 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-bb5b4b448-ql656" podStartSLOduration=1.538914616 podStartE2EDuration="12.760759012s" podCreationTimestamp="2025-12-03 13:21:50 +0000 UTC" firstStartedPulling="2025-12-03 13:21:51.210963007 +0000 UTC m=+757.191883440" lastFinishedPulling="2025-12-03 13:22:02.432807403 +0000 UTC m=+768.413727836" observedRunningTime="2025-12-03 13:22:02.758267799 +0000 UTC m=+768.739188252" watchObservedRunningTime="2025-12-03 13:22:02.760759012 +0000 UTC m=+768.741679435" Dec 03 13:22:07 crc kubenswrapper[4690]: I1203 13:22:07.763937 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 03 13:22:07 crc kubenswrapper[4690]: I1203 13:22:07.765462 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 03 13:22:07 crc kubenswrapper[4690]: I1203 13:22:07.768855 4690 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-5nb25" Dec 03 13:22:07 crc kubenswrapper[4690]: I1203 13:22:07.769049 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 03 13:22:07 crc kubenswrapper[4690]: I1203 13:22:07.774035 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 03 13:22:07 crc kubenswrapper[4690]: I1203 13:22:07.840966 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 03 13:22:07 crc kubenswrapper[4690]: I1203 13:22:07.941514 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-73522fef-c6d6-4d8b-b2ed-7bca6d54a34f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-73522fef-c6d6-4d8b-b2ed-7bca6d54a34f\") pod \"minio\" (UID: \"57bda254-cdfb-496a-b53d-605ec1216610\") " pod="minio-dev/minio" Dec 03 13:22:07 crc kubenswrapper[4690]: I1203 13:22:07.941603 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djvb2\" (UniqueName: \"kubernetes.io/projected/57bda254-cdfb-496a-b53d-605ec1216610-kube-api-access-djvb2\") pod \"minio\" (UID: \"57bda254-cdfb-496a-b53d-605ec1216610\") " pod="minio-dev/minio" Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.043329 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-73522fef-c6d6-4d8b-b2ed-7bca6d54a34f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-73522fef-c6d6-4d8b-b2ed-7bca6d54a34f\") pod \"minio\" (UID: \"57bda254-cdfb-496a-b53d-605ec1216610\") " pod="minio-dev/minio" Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.043404 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djvb2\" (UniqueName: \"kubernetes.io/projected/57bda254-cdfb-496a-b53d-605ec1216610-kube-api-access-djvb2\") pod \"minio\" (UID: \"57bda254-cdfb-496a-b53d-605ec1216610\") " pod="minio-dev/minio" Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.047645 4690 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.047706 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-73522fef-c6d6-4d8b-b2ed-7bca6d54a34f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-73522fef-c6d6-4d8b-b2ed-7bca6d54a34f\") pod \"minio\" (UID: \"57bda254-cdfb-496a-b53d-605ec1216610\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f00b9a527a03cdc7cf33fa08f577e64421d5c8720e795afbb4f32881768f1ae4/globalmount\"" pod="minio-dev/minio" Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.086606 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djvb2\" (UniqueName: \"kubernetes.io/projected/57bda254-cdfb-496a-b53d-605ec1216610-kube-api-access-djvb2\") pod \"minio\" (UID: \"57bda254-cdfb-496a-b53d-605ec1216610\") " pod="minio-dev/minio" Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.087826 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-73522fef-c6d6-4d8b-b2ed-7bca6d54a34f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-73522fef-c6d6-4d8b-b2ed-7bca6d54a34f\") pod \"minio\" (UID: \"57bda254-cdfb-496a-b53d-605ec1216610\") " pod="minio-dev/minio" Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.156091 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.368297 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 03 13:22:08 crc kubenswrapper[4690]: I1203 13:22:08.765190 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"57bda254-cdfb-496a-b53d-605ec1216610","Type":"ContainerStarted","Data":"b3b4f202cb8734188926cc397d570149b15e1e1d46288ce1f1b70c6bad437afe"} Dec 03 13:22:14 crc kubenswrapper[4690]: I1203 13:22:14.814668 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"57bda254-cdfb-496a-b53d-605ec1216610","Type":"ContainerStarted","Data":"176003b32023046583f1c3aa4a14425f81ce035c38c789fd2495e633f5e454c4"} Dec 03 13:22:14 crc kubenswrapper[4690]: I1203 13:22:14.830835 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=5.352680627 podStartE2EDuration="10.830813858s" podCreationTimestamp="2025-12-03 13:22:04 +0000 UTC" firstStartedPulling="2025-12-03 13:22:08.382242363 +0000 UTC m=+774.363162796" lastFinishedPulling="2025-12-03 13:22:13.860375594 +0000 UTC m=+779.841296027" observedRunningTime="2025-12-03 13:22:14.830333006 +0000 UTC m=+780.811253449" watchObservedRunningTime="2025-12-03 13:22:14.830813858 +0000 UTC m=+780.811734291" Dec 03 13:22:16 crc kubenswrapper[4690]: I1203 13:22:16.823739 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:22:16 crc kubenswrapper[4690]: I1203 13:22:16.825089 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.759968 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh"] Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.761381 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.763899 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.772788 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.772800 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.772792 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.772954 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-s82zj" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.777777 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh"] Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.894564 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4eb66c8-ca39-4ba9-8717-a488817c5304-config\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.894625 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrb7w\" (UniqueName: \"kubernetes.io/projected/d4eb66c8-ca39-4ba9-8717-a488817c5304-kube-api-access-zrb7w\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.894818 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.894994 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.895023 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.932791 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-ckhv9"] Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.933683 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.945039 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.945824 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.946301 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.968975 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-ckhv9"] Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.996718 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrb7w\" (UniqueName: \"kubernetes.io/projected/d4eb66c8-ca39-4ba9-8717-a488817c5304-kube-api-access-zrb7w\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.996825 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.996884 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.996907 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.996971 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4eb66c8-ca39-4ba9-8717-a488817c5304-config\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.997964 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:18 crc kubenswrapper[4690]: I1203 13:22:18.998275 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4eb66c8-ca39-4ba9-8717-a488817c5304-config\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.006457 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.024338 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d4eb66c8-ca39-4ba9-8717-a488817c5304-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.027004 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrb7w\" (UniqueName: \"kubernetes.io/projected/d4eb66c8-ca39-4ba9-8717-a488817c5304-kube-api-access-zrb7w\") pod \"logging-loki-distributor-76cc67bf56-qx2hh\" (UID: \"d4eb66c8-ca39-4ba9-8717-a488817c5304\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.081173 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.098125 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.098184 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.098213 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.098378 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.098464 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s2wb\" (UniqueName: \"kubernetes.io/projected/36825bd7-9921-4f76-a26d-6dd05f00bc3e-kube-api-access-9s2wb\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.098759 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36825bd7-9921-4f76-a26d-6dd05f00bc3e-config\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.140949 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.143306 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.147910 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.148189 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.175547 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.201150 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.201212 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.201234 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.201257 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.201281 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s2wb\" (UniqueName: \"kubernetes.io/projected/36825bd7-9921-4f76-a26d-6dd05f00bc3e-kube-api-access-9s2wb\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.201370 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36825bd7-9921-4f76-a26d-6dd05f00bc3e-config\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.202625 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.203013 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36825bd7-9921-4f76-a26d-6dd05f00bc3e-config\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.209708 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.211507 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.223504 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/36825bd7-9921-4f76-a26d-6dd05f00bc3e-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.226275 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s2wb\" (UniqueName: \"kubernetes.io/projected/36825bd7-9921-4f76-a26d-6dd05f00bc3e-kube-api-access-9s2wb\") pod \"logging-loki-querier-5895d59bb8-ckhv9\" (UID: \"36825bd7-9921-4f76-a26d-6dd05f00bc3e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.253535 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.273700 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-79d97f9547-sps84"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.282097 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.288074 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.288140 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.292409 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.292558 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.292671 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.292839 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-bbckb" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.302622 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.302677 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbkw2\" (UniqueName: \"kubernetes.io/projected/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-kube-api-access-kbkw2\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.302715 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-config\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.302748 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.302779 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.304447 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-79d97f9547-8tgpn"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.305407 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.340247 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-79d97f9547-sps84"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.365696 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-79d97f9547-8tgpn"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.407716 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408502 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-lokistack-gateway\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408579 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbkw2\" (UniqueName: \"kubernetes.io/projected/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-kube-api-access-kbkw2\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408607 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-tls-secret\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408648 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-rbac\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408677 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-config\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408731 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-rbac\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408754 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408787 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408810 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-lokistack-gateway\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408882 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mccr4\" (UniqueName: \"kubernetes.io/projected/f1541c44-fb2e-4414-88f0-21e464b629f0-kube-api-access-mccr4\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408916 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408961 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-tls-secret\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.408986 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.409013 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.409046 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.409069 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.409101 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhl7s\" (UniqueName: \"kubernetes.io/projected/4b3976e3-672e-4117-a9c6-9a7f34d42be7-kube-api-access-hhl7s\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.409130 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.409156 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-tenants\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.409200 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-tenants\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.410683 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-config\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.411763 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.420975 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.421081 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.438852 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbkw2\" (UniqueName: \"kubernetes.io/projected/1b8cb794-f99d-41b0-bbdc-ed052be8f4a5-kube-api-access-kbkw2\") pod \"logging-loki-query-frontend-84558f7c9f-pslms\" (UID: \"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.484118 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.510441 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-rbac\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.510930 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-rbac\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.510973 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511010 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-lokistack-gateway\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511047 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mccr4\" (UniqueName: \"kubernetes.io/projected/f1541c44-fb2e-4414-88f0-21e464b629f0-kube-api-access-mccr4\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511076 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-tls-secret\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511097 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511121 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511142 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511162 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511186 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhl7s\" (UniqueName: \"kubernetes.io/projected/4b3976e3-672e-4117-a9c6-9a7f34d42be7-kube-api-access-hhl7s\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511204 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511219 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-tenants\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511244 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-tenants\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511279 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-lokistack-gateway\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511306 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-tls-secret\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.511478 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-rbac\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.512151 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-rbac\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.512734 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.513392 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-lokistack-gateway\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.513411 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.513630 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-lokistack-gateway\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.513952 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.514094 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-ca-bundle\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.516732 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.518365 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-tenants\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.518378 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-tls-secret\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.520049 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-tenants\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.521036 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4b3976e3-672e-4117-a9c6-9a7f34d42be7-tls-secret\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.528210 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f1541c44-fb2e-4414-88f0-21e464b629f0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.537094 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhl7s\" (UniqueName: \"kubernetes.io/projected/4b3976e3-672e-4117-a9c6-9a7f34d42be7-kube-api-access-hhl7s\") pod \"logging-loki-gateway-79d97f9547-8tgpn\" (UID: \"4b3976e3-672e-4117-a9c6-9a7f34d42be7\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.539075 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mccr4\" (UniqueName: \"kubernetes.io/projected/f1541c44-fb2e-4414-88f0-21e464b629f0-kube-api-access-mccr4\") pod \"logging-loki-gateway-79d97f9547-sps84\" (UID: \"f1541c44-fb2e-4414-88f0-21e464b629f0\") " pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.559484 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-ckhv9"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.621828 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.636306 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.723805 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.783239 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.844042 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" event={"ID":"d4eb66c8-ca39-4ba9-8717-a488817c5304","Type":"ContainerStarted","Data":"eb0c74bb64c7016d4c95462eff2fde166834bea56124003206da79f37abc0182"} Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.844937 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" event={"ID":"36825bd7-9921-4f76-a26d-6dd05f00bc3e","Type":"ContainerStarted","Data":"0e37dd6968d4a6e94dde1b31b64156abac8d0f1d4abfb673d51f188988e52197"} Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.846339 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" event={"ID":"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5","Type":"ContainerStarted","Data":"192cb568bb5393f53cb18793d2e42ce5195d1912eaac6b3c34f728124e492eaf"} Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.938880 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-79d97f9547-sps84"] Dec 03 13:22:19 crc kubenswrapper[4690]: W1203 13:22:19.943335 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1541c44_fb2e_4414_88f0_21e464b629f0.slice/crio-a372786745dc0bcf80f9b432fffe268dfb3b7c32ac304cb5d7370a46668ebc8d WatchSource:0}: Error finding container a372786745dc0bcf80f9b432fffe268dfb3b7c32ac304cb5d7370a46668ebc8d: Status 404 returned error can't find the container with id a372786745dc0bcf80f9b432fffe268dfb3b7c32ac304cb5d7370a46668ebc8d Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.976893 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.977740 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.982361 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.982361 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 03 13:22:19 crc kubenswrapper[4690]: I1203 13:22:19.993065 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-79d97f9547-8tgpn"] Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.016583 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.065209 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.068014 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.069845 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.069858 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.076726 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.126120 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-73540916-e435-4401-9b9f-b078e7f7dafa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-73540916-e435-4401-9b9f-b078e7f7dafa\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.126176 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.126200 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f6273c31-c8f3-467d-b292-44ff7f92164c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6273c31-c8f3-467d-b292-44ff7f92164c\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.126294 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-config\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.126344 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.126370 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.126401 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qlp5\" (UniqueName: \"kubernetes.io/projected/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-kube-api-access-6qlp5\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.126428 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.206928 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.208344 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.212955 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.213081 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.222577 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228049 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5c0b2b02-489a-4a57-9a6c-487c4a64b075\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c0b2b02-489a-4a57-9a6c-487c4a64b075\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228111 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-config\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228150 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hchvh\" (UniqueName: \"kubernetes.io/projected/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-kube-api-access-hchvh\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228174 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228196 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228221 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228251 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qlp5\" (UniqueName: \"kubernetes.io/projected/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-kube-api-access-6qlp5\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228276 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-config\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228296 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228314 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228363 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-73540916-e435-4401-9b9f-b078e7f7dafa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-73540916-e435-4401-9b9f-b078e7f7dafa\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228380 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228407 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228426 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.228450 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f6273c31-c8f3-467d-b292-44ff7f92164c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6273c31-c8f3-467d-b292-44ff7f92164c\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.229714 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.229724 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-config\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.234720 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.237482 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.240089 4690 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.240158 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f6273c31-c8f3-467d-b292-44ff7f92164c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6273c31-c8f3-467d-b292-44ff7f92164c\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1bb205918703148b87cd0372abeb5bc28ace9d65b3addb460f3d48de5ae90e60/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.240268 4690 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.240356 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-73540916-e435-4401-9b9f-b078e7f7dafa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-73540916-e435-4401-9b9f-b078e7f7dafa\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/38e3255b32d4712f56f87c450ba4821b091eaa9aa2096c44546abbcb3a001809/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.241358 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.251368 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qlp5\" (UniqueName: \"kubernetes.io/projected/ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86-kube-api-access-6qlp5\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.273233 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f6273c31-c8f3-467d-b292-44ff7f92164c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6273c31-c8f3-467d-b292-44ff7f92164c\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.280260 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-73540916-e435-4401-9b9f-b078e7f7dafa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-73540916-e435-4401-9b9f-b078e7f7dafa\") pod \"logging-loki-ingester-0\" (UID: \"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.294294 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330309 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hchvh\" (UniqueName: \"kubernetes.io/projected/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-kube-api-access-hchvh\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330394 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330438 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330470 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-config\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330496 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330528 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56039ef-9919-4d3d-9b04-876bad0f5756-config\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330573 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-834a5a7b-94ee-4e5f-93df-e171e12c8402\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-834a5a7b-94ee-4e5f-93df-e171e12c8402\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330601 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330640 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb4d6\" (UniqueName: \"kubernetes.io/projected/e56039ef-9919-4d3d-9b04-876bad0f5756-kube-api-access-mb4d6\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330665 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330703 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5c0b2b02-489a-4a57-9a6c-487c4a64b075\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c0b2b02-489a-4a57-9a6c-487c4a64b075\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330738 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330780 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.330814 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.332614 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-config\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.332795 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.335697 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.335886 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.337412 4690 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.337461 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5c0b2b02-489a-4a57-9a6c-487c4a64b075\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c0b2b02-489a-4a57-9a6c-487c4a64b075\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dc1cba2e1269b35aa4fb511ed78bbfde4bc00c798ab1701ecd160c3b6e3cd2be/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.342214 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.352158 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hchvh\" (UniqueName: \"kubernetes.io/projected/6289c341-366c-4b2e-a4ed-2d0eb43d0da4-kube-api-access-hchvh\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.361561 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5c0b2b02-489a-4a57-9a6c-487c4a64b075\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c0b2b02-489a-4a57-9a6c-487c4a64b075\") pod \"logging-loki-compactor-0\" (UID: \"6289c341-366c-4b2e-a4ed-2d0eb43d0da4\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.386510 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.434645 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56039ef-9919-4d3d-9b04-876bad0f5756-config\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.434714 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-834a5a7b-94ee-4e5f-93df-e171e12c8402\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-834a5a7b-94ee-4e5f-93df-e171e12c8402\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.434745 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb4d6\" (UniqueName: \"kubernetes.io/projected/e56039ef-9919-4d3d-9b04-876bad0f5756-kube-api-access-mb4d6\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.434790 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.434817 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.434833 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.434857 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.437383 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.438023 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56039ef-9919-4d3d-9b04-876bad0f5756-config\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.445572 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.450955 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.451706 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/e56039ef-9919-4d3d-9b04-876bad0f5756-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.457497 4690 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.457538 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-834a5a7b-94ee-4e5f-93df-e171e12c8402\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-834a5a7b-94ee-4e5f-93df-e171e12c8402\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/85bb73009dca007ec75b2e39715040455cf3b8e613b9aff8f4e79a597e71db70/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.460722 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb4d6\" (UniqueName: \"kubernetes.io/projected/e56039ef-9919-4d3d-9b04-876bad0f5756-kube-api-access-mb4d6\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.570384 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.591975 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-834a5a7b-94ee-4e5f-93df-e171e12c8402\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-834a5a7b-94ee-4e5f-93df-e171e12c8402\") pod \"logging-loki-index-gateway-0\" (UID: \"e56039ef-9919-4d3d-9b04-876bad0f5756\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.744136 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 13:22:20 crc kubenswrapper[4690]: W1203 13:22:20.744893 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6289c341_366c_4b2e_a4ed_2d0eb43d0da4.slice/crio-34a5baa07f5f706f9834a86de13033ced9c89f389fdaa23fdfd6205527710a35 WatchSource:0}: Error finding container 34a5baa07f5f706f9834a86de13033ced9c89f389fdaa23fdfd6205527710a35: Status 404 returned error can't find the container with id 34a5baa07f5f706f9834a86de13033ced9c89f389fdaa23fdfd6205527710a35 Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.837395 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.854224 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" event={"ID":"4b3976e3-672e-4117-a9c6-9a7f34d42be7","Type":"ContainerStarted","Data":"a8d2f45333e8f37e49fcae74bc123e48a4a5c18ef8d4484e5c9ec5d0b5e9fe18"} Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.855382 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" event={"ID":"f1541c44-fb2e-4414-88f0-21e464b629f0","Type":"ContainerStarted","Data":"a372786745dc0bcf80f9b432fffe268dfb3b7c32ac304cb5d7370a46668ebc8d"} Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.856813 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"6289c341-366c-4b2e-a4ed-2d0eb43d0da4","Type":"ContainerStarted","Data":"34a5baa07f5f706f9834a86de13033ced9c89f389fdaa23fdfd6205527710a35"} Dec 03 13:22:20 crc kubenswrapper[4690]: I1203 13:22:20.858183 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86","Type":"ContainerStarted","Data":"dec9bd098a37b7108e81a5fb30852a95e98c84fe8d5c96631af2019e88c78487"} Dec 03 13:22:21 crc kubenswrapper[4690]: I1203 13:22:21.022514 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 13:22:21 crc kubenswrapper[4690]: I1203 13:22:21.871751 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"e56039ef-9919-4d3d-9b04-876bad0f5756","Type":"ContainerStarted","Data":"0465b0efb4db02072805abb5571767897c66dcec1230eb7b318da55f5fd78851"} Dec 03 13:22:23 crc kubenswrapper[4690]: I1203 13:22:23.886604 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" event={"ID":"f1541c44-fb2e-4414-88f0-21e464b629f0","Type":"ContainerStarted","Data":"5edcd4e0b4a68c37497ea95f652b6a87133ac65fa4b49e1b97f0ad872aa64b07"} Dec 03 13:22:23 crc kubenswrapper[4690]: I1203 13:22:23.888126 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" event={"ID":"4b3976e3-672e-4117-a9c6-9a7f34d42be7","Type":"ContainerStarted","Data":"d836eb7f17830dc099ccf3574aafbd0d45f60d17223cd6ead8ef1655af327e44"} Dec 03 13:22:24 crc kubenswrapper[4690]: I1203 13:22:24.905217 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"e56039ef-9919-4d3d-9b04-876bad0f5756","Type":"ContainerStarted","Data":"4d10cff10d262e86cba70f31f270b9bb34f05591ade5c5ece3aad7f54688fb50"} Dec 03 13:22:24 crc kubenswrapper[4690]: I1203 13:22:24.906050 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:24 crc kubenswrapper[4690]: I1203 13:22:24.917410 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86","Type":"ContainerStarted","Data":"34b2d23eb4b3d0d85e6e48cd1c5ad480aef87101dfdfb0f7ddef6db0bb89297f"} Dec 03 13:22:24 crc kubenswrapper[4690]: I1203 13:22:24.917508 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:22:24 crc kubenswrapper[4690]: I1203 13:22:24.924558 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:24 crc kubenswrapper[4690]: I1203 13:22:24.931920 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=2.377976279 podStartE2EDuration="5.931905393s" podCreationTimestamp="2025-12-03 13:22:19 +0000 UTC" firstStartedPulling="2025-12-03 13:22:21.043148575 +0000 UTC m=+787.024069008" lastFinishedPulling="2025-12-03 13:22:24.597077649 +0000 UTC m=+790.577998122" observedRunningTime="2025-12-03 13:22:24.927532892 +0000 UTC m=+790.908453325" watchObservedRunningTime="2025-12-03 13:22:24.931905393 +0000 UTC m=+790.912825816" Dec 03 13:22:24 crc kubenswrapper[4690]: I1203 13:22:24.971073 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=2.8594181499999998 podStartE2EDuration="6.971056917s" podCreationTimestamp="2025-12-03 13:22:18 +0000 UTC" firstStartedPulling="2025-12-03 13:22:20.593497496 +0000 UTC m=+786.574417929" lastFinishedPulling="2025-12-03 13:22:24.705136263 +0000 UTC m=+790.686056696" observedRunningTime="2025-12-03 13:22:24.969101927 +0000 UTC m=+790.950022350" watchObservedRunningTime="2025-12-03 13:22:24.971056917 +0000 UTC m=+790.951977340" Dec 03 13:22:24 crc kubenswrapper[4690]: I1203 13:22:24.971669 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" podStartSLOduration=1.604616628 podStartE2EDuration="5.971664392s" podCreationTimestamp="2025-12-03 13:22:19 +0000 UTC" firstStartedPulling="2025-12-03 13:22:19.791406346 +0000 UTC m=+785.772326779" lastFinishedPulling="2025-12-03 13:22:24.15845411 +0000 UTC m=+790.139374543" observedRunningTime="2025-12-03 13:22:24.95347164 +0000 UTC m=+790.934392093" watchObservedRunningTime="2025-12-03 13:22:24.971664392 +0000 UTC m=+790.952584825" Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.932915 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" event={"ID":"1b8cb794-f99d-41b0-bbdc-ed052be8f4a5","Type":"ContainerStarted","Data":"ff562b3920b5402f2fddbfa313aa93738f33b3839d7396e6708718e1a3e82fbc"} Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.935767 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" event={"ID":"d4eb66c8-ca39-4ba9-8717-a488817c5304","Type":"ContainerStarted","Data":"e1ebcb66ea213d14803c27f6230038dae11eb4f7a7f6e73d1aa2fa54841d8bf6"} Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.935911 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.937118 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" event={"ID":"36825bd7-9921-4f76-a26d-6dd05f00bc3e","Type":"ContainerStarted","Data":"ba1d20d57754558d0f842ee3b9a9f3076dabb019c8b59b49e8875430cc0fbf60"} Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.937253 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.940563 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"6289c341-366c-4b2e-a4ed-2d0eb43d0da4","Type":"ContainerStarted","Data":"a6692b5403bb1a54cdba8a7ab056fbc3dbce92abacb49cb09124b8285e5d58c5"} Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.963742 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" podStartSLOduration=2.94255277 podStartE2EDuration="7.963713805s" podCreationTimestamp="2025-12-03 13:22:18 +0000 UTC" firstStartedPulling="2025-12-03 13:22:19.752446287 +0000 UTC m=+785.733366720" lastFinishedPulling="2025-12-03 13:22:24.773607322 +0000 UTC m=+790.754527755" observedRunningTime="2025-12-03 13:22:25.953183948 +0000 UTC m=+791.934104401" watchObservedRunningTime="2025-12-03 13:22:25.963713805 +0000 UTC m=+791.944634238" Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.983907 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.119321885 podStartE2EDuration="6.983835176s" podCreationTimestamp="2025-12-03 13:22:19 +0000 UTC" firstStartedPulling="2025-12-03 13:22:20.747452156 +0000 UTC m=+786.728372589" lastFinishedPulling="2025-12-03 13:22:24.611965427 +0000 UTC m=+790.592885880" observedRunningTime="2025-12-03 13:22:25.971278247 +0000 UTC m=+791.952198690" watchObservedRunningTime="2025-12-03 13:22:25.983835176 +0000 UTC m=+791.964755599" Dec 03 13:22:25 crc kubenswrapper[4690]: I1203 13:22:25.989362 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" podStartSLOduration=2.494953732 podStartE2EDuration="7.989344386s" podCreationTimestamp="2025-12-03 13:22:18 +0000 UTC" firstStartedPulling="2025-12-03 13:22:19.587665522 +0000 UTC m=+785.568585955" lastFinishedPulling="2025-12-03 13:22:25.082056176 +0000 UTC m=+791.062976609" observedRunningTime="2025-12-03 13:22:25.987718015 +0000 UTC m=+791.968638448" watchObservedRunningTime="2025-12-03 13:22:25.989344386 +0000 UTC m=+791.970264819" Dec 03 13:22:26 crc kubenswrapper[4690]: I1203 13:22:26.949162 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" event={"ID":"4b3976e3-672e-4117-a9c6-9a7f34d42be7","Type":"ContainerStarted","Data":"cd87f16a385ba446b84d5e76675fcf4ae4c0cc90f155331e4d20fb08d1d70ce7"} Dec 03 13:22:26 crc kubenswrapper[4690]: I1203 13:22:26.949442 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:26 crc kubenswrapper[4690]: I1203 13:22:26.951934 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" event={"ID":"f1541c44-fb2e-4414-88f0-21e464b629f0","Type":"ContainerStarted","Data":"c4ada66d60bf196ec71ccb1c56a8593d69d24cc218a8c4f232118208a5d9cdcd"} Dec 03 13:22:26 crc kubenswrapper[4690]: I1203 13:22:26.952627 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:26 crc kubenswrapper[4690]: I1203 13:22:26.962363 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:26 crc kubenswrapper[4690]: I1203 13:22:26.969890 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" podStartSLOduration=2.099105926 podStartE2EDuration="7.969858557s" podCreationTimestamp="2025-12-03 13:22:19 +0000 UTC" firstStartedPulling="2025-12-03 13:22:20.007060363 +0000 UTC m=+785.987980796" lastFinishedPulling="2025-12-03 13:22:25.877812994 +0000 UTC m=+791.858733427" observedRunningTime="2025-12-03 13:22:26.965610289 +0000 UTC m=+792.946530722" watchObservedRunningTime="2025-12-03 13:22:26.969858557 +0000 UTC m=+792.950778990" Dec 03 13:22:27 crc kubenswrapper[4690]: I1203 13:22:27.010247 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" podStartSLOduration=2.07131157 podStartE2EDuration="8.010228052s" podCreationTimestamp="2025-12-03 13:22:19 +0000 UTC" firstStartedPulling="2025-12-03 13:22:19.945727055 +0000 UTC m=+785.926647488" lastFinishedPulling="2025-12-03 13:22:25.884643537 +0000 UTC m=+791.865563970" observedRunningTime="2025-12-03 13:22:27.005822091 +0000 UTC m=+792.986742524" watchObservedRunningTime="2025-12-03 13:22:27.010228052 +0000 UTC m=+792.991148485" Dec 03 13:22:28 crc kubenswrapper[4690]: I1203 13:22:28.000055 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:28 crc kubenswrapper[4690]: I1203 13:22:28.000098 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:28 crc kubenswrapper[4690]: I1203 13:22:28.000110 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:28 crc kubenswrapper[4690]: I1203 13:22:28.007410 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:28 crc kubenswrapper[4690]: I1203 13:22:28.013311 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-79d97f9547-sps84" Dec 03 13:22:28 crc kubenswrapper[4690]: I1203 13:22:28.014814 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-79d97f9547-8tgpn" Dec 03 13:22:39 crc kubenswrapper[4690]: I1203 13:22:39.491716 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-pslms" Dec 03 13:22:40 crc kubenswrapper[4690]: I1203 13:22:40.305725 4690 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 03 13:22:40 crc kubenswrapper[4690]: I1203 13:22:40.305820 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 13:22:40 crc kubenswrapper[4690]: I1203 13:22:40.395927 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 03 13:22:40 crc kubenswrapper[4690]: I1203 13:22:40.845531 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 13:22:46 crc kubenswrapper[4690]: I1203 13:22:46.823982 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:22:46 crc kubenswrapper[4690]: I1203 13:22:46.824700 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:22:49 crc kubenswrapper[4690]: I1203 13:22:49.086511 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-qx2hh" Dec 03 13:22:49 crc kubenswrapper[4690]: I1203 13:22:49.266603 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-ckhv9" Dec 03 13:22:50 crc kubenswrapper[4690]: I1203 13:22:50.305389 4690 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 03 13:22:50 crc kubenswrapper[4690]: I1203 13:22:50.305468 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 13:23:00 crc kubenswrapper[4690]: I1203 13:23:00.303236 4690 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 03 13:23:00 crc kubenswrapper[4690]: I1203 13:23:00.303615 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 13:23:10 crc kubenswrapper[4690]: I1203 13:23:10.303917 4690 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 03 13:23:10 crc kubenswrapper[4690]: I1203 13:23:10.304766 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 13:23:16 crc kubenswrapper[4690]: I1203 13:23:16.823824 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:23:16 crc kubenswrapper[4690]: I1203 13:23:16.824736 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:23:16 crc kubenswrapper[4690]: I1203 13:23:16.824795 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:23:16 crc kubenswrapper[4690]: I1203 13:23:16.825683 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6757d3773013b08b3fe1dc029582fe576a1dfe2c310afb7b61e2a8c64d0aa40f"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:23:16 crc kubenswrapper[4690]: I1203 13:23:16.825743 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://6757d3773013b08b3fe1dc029582fe576a1dfe2c310afb7b61e2a8c64d0aa40f" gracePeriod=600 Dec 03 13:23:17 crc kubenswrapper[4690]: I1203 13:23:17.351809 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="6757d3773013b08b3fe1dc029582fe576a1dfe2c310afb7b61e2a8c64d0aa40f" exitCode=0 Dec 03 13:23:17 crc kubenswrapper[4690]: I1203 13:23:17.352235 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"6757d3773013b08b3fe1dc029582fe576a1dfe2c310afb7b61e2a8c64d0aa40f"} Dec 03 13:23:17 crc kubenswrapper[4690]: I1203 13:23:17.352554 4690 scope.go:117] "RemoveContainer" containerID="c69bb0d7919080bb99202ef400a31cbfc36de1e7353348ec9eea0dfce4422b6f" Dec 03 13:23:18 crc kubenswrapper[4690]: I1203 13:23:18.363854 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"7081acf7649c775529e892a032f5a109816cdd926c9cbd0790c89b5ae1594b9f"} Dec 03 13:23:20 crc kubenswrapper[4690]: I1203 13:23:20.305512 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.894703 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-g4jxm"] Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.896332 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-g4jxm" Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.919605 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.919890 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.920040 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.920291 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-g4jxm"] Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.923418 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-drjfv" Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.927260 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 03 13:23:39 crc kubenswrapper[4690]: I1203 13:23:39.927952 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.043356 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-g4jxm"] Dec 03 13:23:40 crc kubenswrapper[4690]: E1203 13:23:40.043857 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-wgqf7 metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-g4jxm" podUID="252b4175-af00-4a8a-9327-0111ad819161" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052056 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052106 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-trusted-ca\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052127 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-entrypoint\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052145 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config-openshift-service-cacrt\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052167 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/252b4175-af00-4a8a-9327-0111ad819161-datadir\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052188 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052208 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-sa-token\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052315 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-token\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052486 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgqf7\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-kube-api-access-wgqf7\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052551 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/252b4175-af00-4a8a-9327-0111ad819161-tmp\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.052582 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-syslog-receiver\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154127 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-sa-token\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154208 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-token\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154279 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgqf7\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-kube-api-access-wgqf7\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154317 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/252b4175-af00-4a8a-9327-0111ad819161-tmp\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154342 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-syslog-receiver\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154376 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154412 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-trusted-ca\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154438 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-entrypoint\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154461 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config-openshift-service-cacrt\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154494 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/252b4175-af00-4a8a-9327-0111ad819161-datadir\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.154520 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: E1203 13:23:40.154580 4690 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Dec 03 13:23:40 crc kubenswrapper[4690]: E1203 13:23:40.154637 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics podName:252b4175-af00-4a8a-9327-0111ad819161 nodeName:}" failed. No retries permitted until 2025-12-03 13:23:40.654621339 +0000 UTC m=+866.635541772 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics") pod "collector-g4jxm" (UID: "252b4175-af00-4a8a-9327-0111ad819161") : secret "collector-metrics" not found Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.156218 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/252b4175-af00-4a8a-9327-0111ad819161-datadir\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.156391 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-trusted-ca\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.156420 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-entrypoint\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.156835 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config-openshift-service-cacrt\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.157506 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.165024 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/252b4175-af00-4a8a-9327-0111ad819161-tmp\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.165418 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-token\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.167756 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-syslog-receiver\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.171185 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgqf7\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-kube-api-access-wgqf7\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.174995 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-sa-token\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.504848 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.515317 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662577 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662664 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-token\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662692 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/252b4175-af00-4a8a-9327-0111ad819161-datadir\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662739 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgqf7\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-kube-api-access-wgqf7\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662789 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-syslog-receiver\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662855 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-sa-token\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662906 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-entrypoint\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662925 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/252b4175-af00-4a8a-9327-0111ad819161-tmp\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.662948 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-trusted-ca\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.663009 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config-openshift-service-cacrt\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.663180 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/252b4175-af00-4a8a-9327-0111ad819161-datadir" (OuterVolumeSpecName: "datadir") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.663341 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.663432 4690 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/252b4175-af00-4a8a-9327-0111ad819161-datadir\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.664147 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.664517 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.664913 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.665149 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config" (OuterVolumeSpecName: "config") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.668503 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics\") pod \"collector-g4jxm\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " pod="openshift-logging/collector-g4jxm" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.668932 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.668972 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-token" (OuterVolumeSpecName: "collector-token") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.669522 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/252b4175-af00-4a8a-9327-0111ad819161-tmp" (OuterVolumeSpecName: "tmp") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.672842 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-sa-token" (OuterVolumeSpecName: "sa-token") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.673173 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-kube-api-access-wgqf7" (OuterVolumeSpecName: "kube-api-access-wgqf7") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "kube-api-access-wgqf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.764320 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics\") pod \"252b4175-af00-4a8a-9327-0111ad819161\" (UID: \"252b4175-af00-4a8a-9327-0111ad819161\") " Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.764972 4690 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.764993 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.765002 4690 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.765013 4690 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/252b4175-af00-4a8a-9327-0111ad819161-tmp\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.765023 4690 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.765035 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252b4175-af00-4a8a-9327-0111ad819161-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.765046 4690 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.765057 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgqf7\" (UniqueName: \"kubernetes.io/projected/252b4175-af00-4a8a-9327-0111ad819161-kube-api-access-wgqf7\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.765066 4690 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.767728 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics" (OuterVolumeSpecName: "metrics") pod "252b4175-af00-4a8a-9327-0111ad819161" (UID: "252b4175-af00-4a8a-9327-0111ad819161"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:23:40 crc kubenswrapper[4690]: I1203 13:23:40.866176 4690 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/252b4175-af00-4a8a-9327-0111ad819161-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.510119 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-g4jxm" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.558226 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-g4jxm"] Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.588926 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-g4jxm"] Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.594082 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-g89tt"] Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.595102 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.599151 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.599991 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.600277 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.600400 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.601771 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-drjfv" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.606611 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-g89tt"] Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.612235 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680402 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-config\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680458 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqc62\" (UniqueName: \"kubernetes.io/projected/ed87f330-d809-4183-8cc7-61caf74208a8-kube-api-access-vqc62\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680498 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-metrics\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680570 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-config-openshift-service-cacrt\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680631 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ed87f330-d809-4183-8cc7-61caf74208a8-tmp\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680674 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/ed87f330-d809-4183-8cc7-61caf74208a8-sa-token\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680712 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-trusted-ca\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680750 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/ed87f330-d809-4183-8cc7-61caf74208a8-datadir\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680773 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-collector-syslog-receiver\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680829 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-entrypoint\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.680919 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-collector-token\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.782679 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqc62\" (UniqueName: \"kubernetes.io/projected/ed87f330-d809-4183-8cc7-61caf74208a8-kube-api-access-vqc62\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.782752 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-metrics\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.782793 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-config-openshift-service-cacrt\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.782836 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ed87f330-d809-4183-8cc7-61caf74208a8-tmp\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.782888 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/ed87f330-d809-4183-8cc7-61caf74208a8-sa-token\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.782919 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-trusted-ca\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.782959 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/ed87f330-d809-4183-8cc7-61caf74208a8-datadir\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.782983 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-collector-syslog-receiver\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.783010 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-entrypoint\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.783049 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-collector-token\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.783085 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-config\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.783157 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/ed87f330-d809-4183-8cc7-61caf74208a8-datadir\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.784035 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-config-openshift-service-cacrt\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.784049 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-entrypoint\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.784225 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-config\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.785352 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed87f330-d809-4183-8cc7-61caf74208a8-trusted-ca\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.787510 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-metrics\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.792122 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-collector-token\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.795431 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/ed87f330-d809-4183-8cc7-61caf74208a8-collector-syslog-receiver\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.796727 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ed87f330-d809-4183-8cc7-61caf74208a8-tmp\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.802317 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/ed87f330-d809-4183-8cc7-61caf74208a8-sa-token\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.805472 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqc62\" (UniqueName: \"kubernetes.io/projected/ed87f330-d809-4183-8cc7-61caf74208a8-kube-api-access-vqc62\") pod \"collector-g89tt\" (UID: \"ed87f330-d809-4183-8cc7-61caf74208a8\") " pod="openshift-logging/collector-g89tt" Dec 03 13:23:41 crc kubenswrapper[4690]: I1203 13:23:41.914561 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-g89tt" Dec 03 13:23:42 crc kubenswrapper[4690]: I1203 13:23:42.326658 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="252b4175-af00-4a8a-9327-0111ad819161" path="/var/lib/kubelet/pods/252b4175-af00-4a8a-9327-0111ad819161/volumes" Dec 03 13:23:42 crc kubenswrapper[4690]: I1203 13:23:42.327615 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-g89tt"] Dec 03 13:23:42 crc kubenswrapper[4690]: I1203 13:23:42.519650 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-g89tt" event={"ID":"ed87f330-d809-4183-8cc7-61caf74208a8","Type":"ContainerStarted","Data":"49c064fd04baeddbe3990a753b7f5a21aab16cde3480a10e65c81f5db4522e78"} Dec 03 13:23:54 crc kubenswrapper[4690]: I1203 13:23:54.596941 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-g89tt" event={"ID":"ed87f330-d809-4183-8cc7-61caf74208a8","Type":"ContainerStarted","Data":"c3f63078454792daefdd39b71ced89cdf0c54529675cd75d8360a6af70d004ef"} Dec 03 13:23:54 crc kubenswrapper[4690]: I1203 13:23:54.617919 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-g89tt" podStartSLOduration=2.301922403 podStartE2EDuration="13.617899509s" podCreationTimestamp="2025-12-03 13:23:41 +0000 UTC" firstStartedPulling="2025-12-03 13:23:42.331790115 +0000 UTC m=+868.312710558" lastFinishedPulling="2025-12-03 13:23:53.647767231 +0000 UTC m=+879.628687664" observedRunningTime="2025-12-03 13:23:54.616822941 +0000 UTC m=+880.597743394" watchObservedRunningTime="2025-12-03 13:23:54.617899509 +0000 UTC m=+880.598819942" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.699688 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vdl28"] Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.703655 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.711095 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vdl28"] Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.805998 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-utilities\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.806103 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-catalog-content\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.806149 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5589\" (UniqueName: \"kubernetes.io/projected/88a686da-77d0-4d98-b863-45988ca83b70-kube-api-access-j5589\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.907287 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-catalog-content\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.907984 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-catalog-content\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.908145 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5589\" (UniqueName: \"kubernetes.io/projected/88a686da-77d0-4d98-b863-45988ca83b70-kube-api-access-j5589\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.908684 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-utilities\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.909088 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-utilities\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:04 crc kubenswrapper[4690]: I1203 13:24:04.929574 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5589\" (UniqueName: \"kubernetes.io/projected/88a686da-77d0-4d98-b863-45988ca83b70-kube-api-access-j5589\") pod \"community-operators-vdl28\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:05 crc kubenswrapper[4690]: I1203 13:24:05.041664 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:05 crc kubenswrapper[4690]: I1203 13:24:05.376081 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vdl28"] Dec 03 13:24:05 crc kubenswrapper[4690]: I1203 13:24:05.670646 4690 generic.go:334] "Generic (PLEG): container finished" podID="88a686da-77d0-4d98-b863-45988ca83b70" containerID="a3bdc74c5de9d93282b405d208bedbffdaa58d0c44f5a7caefaddd9a7823984a" exitCode=0 Dec 03 13:24:05 crc kubenswrapper[4690]: I1203 13:24:05.670724 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdl28" event={"ID":"88a686da-77d0-4d98-b863-45988ca83b70","Type":"ContainerDied","Data":"a3bdc74c5de9d93282b405d208bedbffdaa58d0c44f5a7caefaddd9a7823984a"} Dec 03 13:24:05 crc kubenswrapper[4690]: I1203 13:24:05.671109 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdl28" event={"ID":"88a686da-77d0-4d98-b863-45988ca83b70","Type":"ContainerStarted","Data":"06b88ea7e07f1e9eae0e46b8845b30bc474710a37af2e9a0fe21439597edb7fe"} Dec 03 13:24:06 crc kubenswrapper[4690]: I1203 13:24:06.680759 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdl28" event={"ID":"88a686da-77d0-4d98-b863-45988ca83b70","Type":"ContainerStarted","Data":"9197d6198ca5f5bdd8f3b8e454f4d5195634384b8cc07efaf2c15a887a26cb32"} Dec 03 13:24:07 crc kubenswrapper[4690]: I1203 13:24:07.689754 4690 generic.go:334] "Generic (PLEG): container finished" podID="88a686da-77d0-4d98-b863-45988ca83b70" containerID="9197d6198ca5f5bdd8f3b8e454f4d5195634384b8cc07efaf2c15a887a26cb32" exitCode=0 Dec 03 13:24:07 crc kubenswrapper[4690]: I1203 13:24:07.689887 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdl28" event={"ID":"88a686da-77d0-4d98-b863-45988ca83b70","Type":"ContainerDied","Data":"9197d6198ca5f5bdd8f3b8e454f4d5195634384b8cc07efaf2c15a887a26cb32"} Dec 03 13:24:08 crc kubenswrapper[4690]: I1203 13:24:08.702069 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdl28" event={"ID":"88a686da-77d0-4d98-b863-45988ca83b70","Type":"ContainerStarted","Data":"0b6c6c26ae33954939041993eee8f08864e1e25ca8fa43d902685443ac2d2cfe"} Dec 03 13:24:08 crc kubenswrapper[4690]: I1203 13:24:08.722833 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vdl28" podStartSLOduration=2.063820855 podStartE2EDuration="4.722807787s" podCreationTimestamp="2025-12-03 13:24:04 +0000 UTC" firstStartedPulling="2025-12-03 13:24:05.672930598 +0000 UTC m=+891.653851031" lastFinishedPulling="2025-12-03 13:24:08.33191753 +0000 UTC m=+894.312837963" observedRunningTime="2025-12-03 13:24:08.718967009 +0000 UTC m=+894.699887452" watchObservedRunningTime="2025-12-03 13:24:08.722807787 +0000 UTC m=+894.703728220" Dec 03 13:24:13 crc kubenswrapper[4690]: I1203 13:24:13.836303 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl"] Dec 03 13:24:13 crc kubenswrapper[4690]: I1203 13:24:13.838238 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:13 crc kubenswrapper[4690]: I1203 13:24:13.840639 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 13:24:13 crc kubenswrapper[4690]: I1203 13:24:13.855005 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl"] Dec 03 13:24:13 crc kubenswrapper[4690]: I1203 13:24:13.951034 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4rhb\" (UniqueName: \"kubernetes.io/projected/314999b8-63d0-4f07-8b80-2aa04364cb5a-kube-api-access-n4rhb\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:13 crc kubenswrapper[4690]: I1203 13:24:13.951101 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:13 crc kubenswrapper[4690]: I1203 13:24:13.951219 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.052230 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.052324 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.052398 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4rhb\" (UniqueName: \"kubernetes.io/projected/314999b8-63d0-4f07-8b80-2aa04364cb5a-kube-api-access-n4rhb\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.052839 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.052846 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.073794 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4rhb\" (UniqueName: \"kubernetes.io/projected/314999b8-63d0-4f07-8b80-2aa04364cb5a-kube-api-access-n4rhb\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.160155 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.585767 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl"] Dec 03 13:24:14 crc kubenswrapper[4690]: I1203 13:24:14.738392 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" event={"ID":"314999b8-63d0-4f07-8b80-2aa04364cb5a","Type":"ContainerStarted","Data":"64bc3113ae02c9ac66fb8ec892e1540d8a84c3ce875949720573bf90dc3e4e0a"} Dec 03 13:24:15 crc kubenswrapper[4690]: I1203 13:24:15.042637 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:15 crc kubenswrapper[4690]: I1203 13:24:15.043981 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:15 crc kubenswrapper[4690]: I1203 13:24:15.086399 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:15 crc kubenswrapper[4690]: I1203 13:24:15.747604 4690 generic.go:334] "Generic (PLEG): container finished" podID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerID="9802aedc221676ce2302f7a18ce155667433555f465fef8f28fdf7dddab8079f" exitCode=0 Dec 03 13:24:15 crc kubenswrapper[4690]: I1203 13:24:15.747673 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" event={"ID":"314999b8-63d0-4f07-8b80-2aa04364cb5a","Type":"ContainerDied","Data":"9802aedc221676ce2302f7a18ce155667433555f465fef8f28fdf7dddab8079f"} Dec 03 13:24:15 crc kubenswrapper[4690]: I1203 13:24:15.795886 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.592953 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9hgpv"] Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.594226 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.606302 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xztz5\" (UniqueName: \"kubernetes.io/projected/49d3e16a-82a0-4cc7-aadc-75496e4ab001-kube-api-access-xztz5\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.606356 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-utilities\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.606405 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-catalog-content\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.607099 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9hgpv"] Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.707105 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xztz5\" (UniqueName: \"kubernetes.io/projected/49d3e16a-82a0-4cc7-aadc-75496e4ab001-kube-api-access-xztz5\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.707156 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-utilities\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.707184 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-catalog-content\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.707638 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-catalog-content\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.707847 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-utilities\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.728102 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xztz5\" (UniqueName: \"kubernetes.io/projected/49d3e16a-82a0-4cc7-aadc-75496e4ab001-kube-api-access-xztz5\") pod \"redhat-operators-9hgpv\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:16 crc kubenswrapper[4690]: I1203 13:24:16.912784 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:17 crc kubenswrapper[4690]: I1203 13:24:17.213662 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9hgpv"] Dec 03 13:24:17 crc kubenswrapper[4690]: I1203 13:24:17.761931 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hgpv" event={"ID":"49d3e16a-82a0-4cc7-aadc-75496e4ab001","Type":"ContainerStarted","Data":"283ffd4e36922c142c714bb62daab9984539f6117131e0bea5541f137f38b715"} Dec 03 13:24:18 crc kubenswrapper[4690]: I1203 13:24:18.585610 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vdl28"] Dec 03 13:24:18 crc kubenswrapper[4690]: I1203 13:24:18.585906 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vdl28" podUID="88a686da-77d0-4d98-b863-45988ca83b70" containerName="registry-server" containerID="cri-o://0b6c6c26ae33954939041993eee8f08864e1e25ca8fa43d902685443ac2d2cfe" gracePeriod=2 Dec 03 13:24:21 crc kubenswrapper[4690]: I1203 13:24:21.795181 4690 generic.go:334] "Generic (PLEG): container finished" podID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerID="7e0cd575101f422989411e41ec37f80eb3d357a160ac14d82389f26c6a81216f" exitCode=0 Dec 03 13:24:21 crc kubenswrapper[4690]: I1203 13:24:21.795301 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hgpv" event={"ID":"49d3e16a-82a0-4cc7-aadc-75496e4ab001","Type":"ContainerDied","Data":"7e0cd575101f422989411e41ec37f80eb3d357a160ac14d82389f26c6a81216f"} Dec 03 13:24:21 crc kubenswrapper[4690]: I1203 13:24:21.799315 4690 generic.go:334] "Generic (PLEG): container finished" podID="88a686da-77d0-4d98-b863-45988ca83b70" containerID="0b6c6c26ae33954939041993eee8f08864e1e25ca8fa43d902685443ac2d2cfe" exitCode=0 Dec 03 13:24:21 crc kubenswrapper[4690]: I1203 13:24:21.799388 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdl28" event={"ID":"88a686da-77d0-4d98-b863-45988ca83b70","Type":"ContainerDied","Data":"0b6c6c26ae33954939041993eee8f08864e1e25ca8fa43d902685443ac2d2cfe"} Dec 03 13:24:22 crc kubenswrapper[4690]: I1203 13:24:22.934761 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.102314 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-catalog-content\") pod \"88a686da-77d0-4d98-b863-45988ca83b70\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.102451 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5589\" (UniqueName: \"kubernetes.io/projected/88a686da-77d0-4d98-b863-45988ca83b70-kube-api-access-j5589\") pod \"88a686da-77d0-4d98-b863-45988ca83b70\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.102511 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-utilities\") pod \"88a686da-77d0-4d98-b863-45988ca83b70\" (UID: \"88a686da-77d0-4d98-b863-45988ca83b70\") " Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.103479 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-utilities" (OuterVolumeSpecName: "utilities") pod "88a686da-77d0-4d98-b863-45988ca83b70" (UID: "88a686da-77d0-4d98-b863-45988ca83b70"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.109090 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a686da-77d0-4d98-b863-45988ca83b70-kube-api-access-j5589" (OuterVolumeSpecName: "kube-api-access-j5589") pod "88a686da-77d0-4d98-b863-45988ca83b70" (UID: "88a686da-77d0-4d98-b863-45988ca83b70"). InnerVolumeSpecName "kube-api-access-j5589". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.153286 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88a686da-77d0-4d98-b863-45988ca83b70" (UID: "88a686da-77d0-4d98-b863-45988ca83b70"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.204498 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.204540 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5589\" (UniqueName: \"kubernetes.io/projected/88a686da-77d0-4d98-b863-45988ca83b70-kube-api-access-j5589\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.204553 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a686da-77d0-4d98-b863-45988ca83b70-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.816059 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdl28" event={"ID":"88a686da-77d0-4d98-b863-45988ca83b70","Type":"ContainerDied","Data":"06b88ea7e07f1e9eae0e46b8845b30bc474710a37af2e9a0fe21439597edb7fe"} Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.816078 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdl28" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.816139 4690 scope.go:117] "RemoveContainer" containerID="0b6c6c26ae33954939041993eee8f08864e1e25ca8fa43d902685443ac2d2cfe" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.820081 4690 generic.go:334] "Generic (PLEG): container finished" podID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerID="93efa2fec7d4e264cb901afd5ddab73ef6b4f20629f8331f5a1573a70f2f0f48" exitCode=0 Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.820134 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" event={"ID":"314999b8-63d0-4f07-8b80-2aa04364cb5a","Type":"ContainerDied","Data":"93efa2fec7d4e264cb901afd5ddab73ef6b4f20629f8331f5a1573a70f2f0f48"} Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.877852 4690 scope.go:117] "RemoveContainer" containerID="9197d6198ca5f5bdd8f3b8e454f4d5195634384b8cc07efaf2c15a887a26cb32" Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.889326 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vdl28"] Dec 03 13:24:23 crc kubenswrapper[4690]: I1203 13:24:23.904765 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vdl28"] Dec 03 13:24:24 crc kubenswrapper[4690]: I1203 13:24:24.063128 4690 scope.go:117] "RemoveContainer" containerID="a3bdc74c5de9d93282b405d208bedbffdaa58d0c44f5a7caefaddd9a7823984a" Dec 03 13:24:24 crc kubenswrapper[4690]: I1203 13:24:24.325887 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a686da-77d0-4d98-b863-45988ca83b70" path="/var/lib/kubelet/pods/88a686da-77d0-4d98-b863-45988ca83b70/volumes" Dec 03 13:24:24 crc kubenswrapper[4690]: I1203 13:24:24.831987 4690 generic.go:334] "Generic (PLEG): container finished" podID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerID="aa4b5aa0c82cc45d48313e46cf0628a24b2887358821357eeeadee875f5d73ac" exitCode=0 Dec 03 13:24:24 crc kubenswrapper[4690]: I1203 13:24:24.832076 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" event={"ID":"314999b8-63d0-4f07-8b80-2aa04364cb5a","Type":"ContainerDied","Data":"aa4b5aa0c82cc45d48313e46cf0628a24b2887358821357eeeadee875f5d73ac"} Dec 03 13:24:24 crc kubenswrapper[4690]: I1203 13:24:24.835315 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hgpv" event={"ID":"49d3e16a-82a0-4cc7-aadc-75496e4ab001","Type":"ContainerStarted","Data":"db9a40adb2499422f710936d9dd62c555817376427acf5b2e5100189079072f1"} Dec 03 13:24:25 crc kubenswrapper[4690]: I1203 13:24:25.845299 4690 generic.go:334] "Generic (PLEG): container finished" podID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerID="db9a40adb2499422f710936d9dd62c555817376427acf5b2e5100189079072f1" exitCode=0 Dec 03 13:24:25 crc kubenswrapper[4690]: I1203 13:24:25.845393 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hgpv" event={"ID":"49d3e16a-82a0-4cc7-aadc-75496e4ab001","Type":"ContainerDied","Data":"db9a40adb2499422f710936d9dd62c555817376427acf5b2e5100189079072f1"} Dec 03 13:24:25 crc kubenswrapper[4690]: I1203 13:24:25.845858 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hgpv" event={"ID":"49d3e16a-82a0-4cc7-aadc-75496e4ab001","Type":"ContainerStarted","Data":"839c5e04508699be9e21711763f4f80c9dc87155a8e4f74e1ff6c97539bc7a24"} Dec 03 13:24:25 crc kubenswrapper[4690]: I1203 13:24:25.885451 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9hgpv" podStartSLOduration=7.387350606 podStartE2EDuration="9.885428514s" podCreationTimestamp="2025-12-03 13:24:16 +0000 UTC" firstStartedPulling="2025-12-03 13:24:22.866732511 +0000 UTC m=+908.847652944" lastFinishedPulling="2025-12-03 13:24:25.364810419 +0000 UTC m=+911.345730852" observedRunningTime="2025-12-03 13:24:25.877738618 +0000 UTC m=+911.858659061" watchObservedRunningTime="2025-12-03 13:24:25.885428514 +0000 UTC m=+911.866348947" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.234301 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.245935 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4rhb\" (UniqueName: \"kubernetes.io/projected/314999b8-63d0-4f07-8b80-2aa04364cb5a-kube-api-access-n4rhb\") pod \"314999b8-63d0-4f07-8b80-2aa04364cb5a\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.246000 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-bundle\") pod \"314999b8-63d0-4f07-8b80-2aa04364cb5a\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.246060 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-util\") pod \"314999b8-63d0-4f07-8b80-2aa04364cb5a\" (UID: \"314999b8-63d0-4f07-8b80-2aa04364cb5a\") " Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.246846 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-bundle" (OuterVolumeSpecName: "bundle") pod "314999b8-63d0-4f07-8b80-2aa04364cb5a" (UID: "314999b8-63d0-4f07-8b80-2aa04364cb5a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.250614 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.255934 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314999b8-63d0-4f07-8b80-2aa04364cb5a-kube-api-access-n4rhb" (OuterVolumeSpecName: "kube-api-access-n4rhb") pod "314999b8-63d0-4f07-8b80-2aa04364cb5a" (UID: "314999b8-63d0-4f07-8b80-2aa04364cb5a"). InnerVolumeSpecName "kube-api-access-n4rhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.266596 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-util" (OuterVolumeSpecName: "util") pod "314999b8-63d0-4f07-8b80-2aa04364cb5a" (UID: "314999b8-63d0-4f07-8b80-2aa04364cb5a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.352784 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314999b8-63d0-4f07-8b80-2aa04364cb5a-util\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.352814 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4rhb\" (UniqueName: \"kubernetes.io/projected/314999b8-63d0-4f07-8b80-2aa04364cb5a-kube-api-access-n4rhb\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.855381 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" event={"ID":"314999b8-63d0-4f07-8b80-2aa04364cb5a","Type":"ContainerDied","Data":"64bc3113ae02c9ac66fb8ec892e1540d8a84c3ce875949720573bf90dc3e4e0a"} Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.856217 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64bc3113ae02c9ac66fb8ec892e1540d8a84c3ce875949720573bf90dc3e4e0a" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.855434 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.913925 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:26 crc kubenswrapper[4690]: I1203 13:24:26.914017 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:27 crc kubenswrapper[4690]: I1203 13:24:27.956027 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9hgpv" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="registry-server" probeResult="failure" output=< Dec 03 13:24:27 crc kubenswrapper[4690]: timeout: failed to connect service ":50051" within 1s Dec 03 13:24:27 crc kubenswrapper[4690]: > Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.082766 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4"] Dec 03 13:24:30 crc kubenswrapper[4690]: E1203 13:24:30.083130 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a686da-77d0-4d98-b863-45988ca83b70" containerName="extract-utilities" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083147 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a686da-77d0-4d98-b863-45988ca83b70" containerName="extract-utilities" Dec 03 13:24:30 crc kubenswrapper[4690]: E1203 13:24:30.083161 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a686da-77d0-4d98-b863-45988ca83b70" containerName="registry-server" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083166 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a686da-77d0-4d98-b863-45988ca83b70" containerName="registry-server" Dec 03 13:24:30 crc kubenswrapper[4690]: E1203 13:24:30.083178 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerName="extract" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083185 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerName="extract" Dec 03 13:24:30 crc kubenswrapper[4690]: E1203 13:24:30.083201 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a686da-77d0-4d98-b863-45988ca83b70" containerName="extract-content" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083208 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a686da-77d0-4d98-b863-45988ca83b70" containerName="extract-content" Dec 03 13:24:30 crc kubenswrapper[4690]: E1203 13:24:30.083218 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerName="util" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083226 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerName="util" Dec 03 13:24:30 crc kubenswrapper[4690]: E1203 13:24:30.083239 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerName="pull" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083246 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerName="pull" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083382 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="314999b8-63d0-4f07-8b80-2aa04364cb5a" containerName="extract" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083403 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a686da-77d0-4d98-b863-45988ca83b70" containerName="registry-server" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.083995 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.086255 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-kpwb7" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.086603 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.086951 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.097644 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4"] Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.109496 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gbwj\" (UniqueName: \"kubernetes.io/projected/0aacb971-67b1-4e6a-a89c-7aa8b404f046-kube-api-access-7gbwj\") pod \"nmstate-operator-5b5b58f5c8-p69d4\" (UID: \"0aacb971-67b1-4e6a-a89c-7aa8b404f046\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.210690 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gbwj\" (UniqueName: \"kubernetes.io/projected/0aacb971-67b1-4e6a-a89c-7aa8b404f046-kube-api-access-7gbwj\") pod \"nmstate-operator-5b5b58f5c8-p69d4\" (UID: \"0aacb971-67b1-4e6a-a89c-7aa8b404f046\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.231094 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gbwj\" (UniqueName: \"kubernetes.io/projected/0aacb971-67b1-4e6a-a89c-7aa8b404f046-kube-api-access-7gbwj\") pod \"nmstate-operator-5b5b58f5c8-p69d4\" (UID: \"0aacb971-67b1-4e6a-a89c-7aa8b404f046\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.407025 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4" Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.847954 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4"] Dec 03 13:24:30 crc kubenswrapper[4690]: W1203 13:24:30.852057 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0aacb971_67b1_4e6a_a89c_7aa8b404f046.slice/crio-be0e0dcc1e60cf1777e7757d126a976733585665fe33cf0e9d8ba981d58dbe36 WatchSource:0}: Error finding container be0e0dcc1e60cf1777e7757d126a976733585665fe33cf0e9d8ba981d58dbe36: Status 404 returned error can't find the container with id be0e0dcc1e60cf1777e7757d126a976733585665fe33cf0e9d8ba981d58dbe36 Dec 03 13:24:30 crc kubenswrapper[4690]: I1203 13:24:30.880066 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4" event={"ID":"0aacb971-67b1-4e6a-a89c-7aa8b404f046","Type":"ContainerStarted","Data":"be0e0dcc1e60cf1777e7757d126a976733585665fe33cf0e9d8ba981d58dbe36"} Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.392909 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4pmnk"] Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.394651 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.404857 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pmnk"] Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.457962 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d44sr\" (UniqueName: \"kubernetes.io/projected/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-kube-api-access-d44sr\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.458077 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-utilities\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.458154 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-catalog-content\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.559671 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d44sr\" (UniqueName: \"kubernetes.io/projected/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-kube-api-access-d44sr\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.559743 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-utilities\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.559803 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-catalog-content\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.560300 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-utilities\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.560338 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-catalog-content\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.585469 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d44sr\" (UniqueName: \"kubernetes.io/projected/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-kube-api-access-d44sr\") pod \"certified-operators-4pmnk\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:33 crc kubenswrapper[4690]: I1203 13:24:33.713312 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:34 crc kubenswrapper[4690]: I1203 13:24:34.227531 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pmnk"] Dec 03 13:24:34 crc kubenswrapper[4690]: I1203 13:24:34.906128 4690 generic.go:334] "Generic (PLEG): container finished" podID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerID="f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869" exitCode=0 Dec 03 13:24:34 crc kubenswrapper[4690]: I1203 13:24:34.906224 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pmnk" event={"ID":"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e","Type":"ContainerDied","Data":"f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869"} Dec 03 13:24:34 crc kubenswrapper[4690]: I1203 13:24:34.906468 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pmnk" event={"ID":"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e","Type":"ContainerStarted","Data":"93b311876337fb7918ef987baac8f8320e289e5282cebe7d7f0c5c56b9444fee"} Dec 03 13:24:36 crc kubenswrapper[4690]: I1203 13:24:36.960944 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:37 crc kubenswrapper[4690]: I1203 13:24:37.009772 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:37 crc kubenswrapper[4690]: I1203 13:24:37.927140 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4" event={"ID":"0aacb971-67b1-4e6a-a89c-7aa8b404f046","Type":"ContainerStarted","Data":"b3c57855e9f50446118cec9e9326814c298bbdaf007647253b7d6a661b622a05"} Dec 03 13:24:37 crc kubenswrapper[4690]: I1203 13:24:37.930988 4690 generic.go:334] "Generic (PLEG): container finished" podID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerID="480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48" exitCode=0 Dec 03 13:24:37 crc kubenswrapper[4690]: I1203 13:24:37.931060 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pmnk" event={"ID":"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e","Type":"ContainerDied","Data":"480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48"} Dec 03 13:24:37 crc kubenswrapper[4690]: I1203 13:24:37.961775 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p69d4" podStartSLOduration=2.986965215 podStartE2EDuration="7.961753257s" podCreationTimestamp="2025-12-03 13:24:30 +0000 UTC" firstStartedPulling="2025-12-03 13:24:30.854874909 +0000 UTC m=+916.835795342" lastFinishedPulling="2025-12-03 13:24:35.829662951 +0000 UTC m=+921.810583384" observedRunningTime="2025-12-03 13:24:37.951175917 +0000 UTC m=+923.932096350" watchObservedRunningTime="2025-12-03 13:24:37.961753257 +0000 UTC m=+923.942673700" Dec 03 13:24:38 crc kubenswrapper[4690]: I1203 13:24:38.958422 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4"] Dec 03 13:24:38 crc kubenswrapper[4690]: I1203 13:24:38.960226 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" Dec 03 13:24:38 crc kubenswrapper[4690]: I1203 13:24:38.963100 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-bxxxh" Dec 03 13:24:38 crc kubenswrapper[4690]: I1203 13:24:38.970705 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4"] Dec 03 13:24:38 crc kubenswrapper[4690]: I1203 13:24:38.990891 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn"] Dec 03 13:24:38 crc kubenswrapper[4690]: I1203 13:24:38.991942 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.003362 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.022064 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-v429n"] Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.022952 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.044119 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn"] Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.131328 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d"] Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.132190 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.134144 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.134323 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.139915 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-62h65" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.141038 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-ovs-socket\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.141107 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-nmstate-lock\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.141330 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6020072e-2f65-4d13-85be-d300b4ec7f6d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-mw7tn\" (UID: \"6020072e-2f65-4d13-85be-d300b4ec7f6d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.141360 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-dbus-socket\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.141464 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzkzb\" (UniqueName: \"kubernetes.io/projected/6020072e-2f65-4d13-85be-d300b4ec7f6d-kube-api-access-kzkzb\") pod \"nmstate-webhook-5f6d4c5ccb-mw7tn\" (UID: \"6020072e-2f65-4d13-85be-d300b4ec7f6d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.141502 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26f2k\" (UniqueName: \"kubernetes.io/projected/89d8054c-b35f-42fd-960d-16e1b39ad983-kube-api-access-26f2k\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.141528 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k92r\" (UniqueName: \"kubernetes.io/projected/33310d90-a18c-44db-a7f3-37f2086de63a-kube-api-access-7k92r\") pod \"nmstate-metrics-7f946cbc9-hjrl4\" (UID: \"33310d90-a18c-44db-a7f3-37f2086de63a\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.175857 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d"] Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.246677 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb3da90d-da37-49ad-ad46-230759602e46-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.246748 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-nmstate-lock\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.246786 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb3da90d-da37-49ad-ad46-230759602e46-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.246821 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6020072e-2f65-4d13-85be-d300b4ec7f6d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-mw7tn\" (UID: \"6020072e-2f65-4d13-85be-d300b4ec7f6d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.246848 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-dbus-socket\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.246876 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-nmstate-lock\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: E1203 13:24:39.246987 4690 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 03 13:24:39 crc kubenswrapper[4690]: E1203 13:24:39.247045 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6020072e-2f65-4d13-85be-d300b4ec7f6d-tls-key-pair podName:6020072e-2f65-4d13-85be-d300b4ec7f6d nodeName:}" failed. No retries permitted until 2025-12-03 13:24:39.74702561 +0000 UTC m=+925.727946033 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/6020072e-2f65-4d13-85be-d300b4ec7f6d-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-mw7tn" (UID: "6020072e-2f65-4d13-85be-d300b4ec7f6d") : secret "openshift-nmstate-webhook" not found Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.247187 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-dbus-socket\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.246891 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qtzf\" (UniqueName: \"kubernetes.io/projected/bb3da90d-da37-49ad-ad46-230759602e46-kube-api-access-4qtzf\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.247338 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzkzb\" (UniqueName: \"kubernetes.io/projected/6020072e-2f65-4d13-85be-d300b4ec7f6d-kube-api-access-kzkzb\") pod \"nmstate-webhook-5f6d4c5ccb-mw7tn\" (UID: \"6020072e-2f65-4d13-85be-d300b4ec7f6d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.247416 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26f2k\" (UniqueName: \"kubernetes.io/projected/89d8054c-b35f-42fd-960d-16e1b39ad983-kube-api-access-26f2k\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.247451 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k92r\" (UniqueName: \"kubernetes.io/projected/33310d90-a18c-44db-a7f3-37f2086de63a-kube-api-access-7k92r\") pod \"nmstate-metrics-7f946cbc9-hjrl4\" (UID: \"33310d90-a18c-44db-a7f3-37f2086de63a\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.247642 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-ovs-socket\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.247733 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/89d8054c-b35f-42fd-960d-16e1b39ad983-ovs-socket\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.285301 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzkzb\" (UniqueName: \"kubernetes.io/projected/6020072e-2f65-4d13-85be-d300b4ec7f6d-kube-api-access-kzkzb\") pod \"nmstate-webhook-5f6d4c5ccb-mw7tn\" (UID: \"6020072e-2f65-4d13-85be-d300b4ec7f6d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.286862 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k92r\" (UniqueName: \"kubernetes.io/projected/33310d90-a18c-44db-a7f3-37f2086de63a-kube-api-access-7k92r\") pod \"nmstate-metrics-7f946cbc9-hjrl4\" (UID: \"33310d90-a18c-44db-a7f3-37f2086de63a\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.301166 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26f2k\" (UniqueName: \"kubernetes.io/projected/89d8054c-b35f-42fd-960d-16e1b39ad983-kube-api-access-26f2k\") pod \"nmstate-handler-v429n\" (UID: \"89d8054c-b35f-42fd-960d-16e1b39ad983\") " pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.345773 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.348727 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb3da90d-da37-49ad-ad46-230759602e46-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.348782 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb3da90d-da37-49ad-ad46-230759602e46-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.348828 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qtzf\" (UniqueName: \"kubernetes.io/projected/bb3da90d-da37-49ad-ad46-230759602e46-kube-api-access-4qtzf\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.350137 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb3da90d-da37-49ad-ad46-230759602e46-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.354109 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-85576f6988-gf7lj"] Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.362172 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.354113 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb3da90d-da37-49ad-ad46-230759602e46-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.388264 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-85576f6988-gf7lj"] Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.396363 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qtzf\" (UniqueName: \"kubernetes.io/projected/bb3da90d-da37-49ad-ad46-230759602e46-kube-api-access-4qtzf\") pod \"nmstate-console-plugin-7fbb5f6569-4xg2d\" (UID: \"bb3da90d-da37-49ad-ad46-230759602e46\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.447977 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.552204 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-oauth-config\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.552267 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-config\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.552335 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-serving-cert\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.552377 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-oauth-serving-cert\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.552446 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-service-ca\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.554021 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6jc2\" (UniqueName: \"kubernetes.io/projected/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-kube-api-access-s6jc2\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.554077 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-trusted-ca-bundle\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.576346 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.660664 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-oauth-config\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.660703 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-config\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.660750 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-serving-cert\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.660776 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-oauth-serving-cert\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.660804 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-service-ca\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.660836 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6jc2\" (UniqueName: \"kubernetes.io/projected/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-kube-api-access-s6jc2\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.660860 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-trusted-ca-bundle\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.661796 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-oauth-serving-cert\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.661921 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-config\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.661949 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-trusted-ca-bundle\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.662641 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-service-ca\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.665313 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-serving-cert\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.665468 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-console-oauth-config\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.684325 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6jc2\" (UniqueName: \"kubernetes.io/projected/35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b-kube-api-access-s6jc2\") pod \"console-85576f6988-gf7lj\" (UID: \"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b\") " pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.741911 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.761818 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6020072e-2f65-4d13-85be-d300b4ec7f6d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-mw7tn\" (UID: \"6020072e-2f65-4d13-85be-d300b4ec7f6d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.764744 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6020072e-2f65-4d13-85be-d300b4ec7f6d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-mw7tn\" (UID: \"6020072e-2f65-4d13-85be-d300b4ec7f6d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.858305 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d"] Dec 03 13:24:39 crc kubenswrapper[4690]: W1203 13:24:39.871025 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb3da90d_da37_49ad_ad46_230759602e46.slice/crio-ea2ca4925a2e0e7dae804de1b79f3fc88c15707bd01d50faab36dfdd356903ed WatchSource:0}: Error finding container ea2ca4925a2e0e7dae804de1b79f3fc88c15707bd01d50faab36dfdd356903ed: Status 404 returned error can't find the container with id ea2ca4925a2e0e7dae804de1b79f3fc88c15707bd01d50faab36dfdd356903ed Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.923224 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.948815 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-85576f6988-gf7lj"] Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.968768 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" event={"ID":"bb3da90d-da37-49ad-ad46-230759602e46","Type":"ContainerStarted","Data":"ea2ca4925a2e0e7dae804de1b79f3fc88c15707bd01d50faab36dfdd356903ed"} Dec 03 13:24:39 crc kubenswrapper[4690]: W1203 13:24:39.970798 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35bd2fae_21a1_4cf8_91fb_08a52c8d9e9b.slice/crio-1ea65eb092f1f5b0dde56794a877be449de87e1c689fbe438e15963ad4ee54a4 WatchSource:0}: Error finding container 1ea65eb092f1f5b0dde56794a877be449de87e1c689fbe438e15963ad4ee54a4: Status 404 returned error can't find the container with id 1ea65eb092f1f5b0dde56794a877be449de87e1c689fbe438e15963ad4ee54a4 Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.973759 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pmnk" event={"ID":"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e","Type":"ContainerStarted","Data":"e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084"} Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.978380 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-v429n" event={"ID":"89d8054c-b35f-42fd-960d-16e1b39ad983","Type":"ContainerStarted","Data":"131a70ee1164721ab1e99fd562e3aedab50896b24729e116910ae548644d0e83"} Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.980233 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4"] Dec 03 13:24:39 crc kubenswrapper[4690]: I1203 13:24:39.995890 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4pmnk" podStartSLOduration=3.072812711 podStartE2EDuration="6.995844904s" podCreationTimestamp="2025-12-03 13:24:33 +0000 UTC" firstStartedPulling="2025-12-03 13:24:34.909706723 +0000 UTC m=+920.890627156" lastFinishedPulling="2025-12-03 13:24:38.832738916 +0000 UTC m=+924.813659349" observedRunningTime="2025-12-03 13:24:39.989301017 +0000 UTC m=+925.970221450" watchObservedRunningTime="2025-12-03 13:24:39.995844904 +0000 UTC m=+925.976765337" Dec 03 13:24:39 crc kubenswrapper[4690]: W1203 13:24:39.996975 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33310d90_a18c_44db_a7f3_37f2086de63a.slice/crio-b9b0d534d481eba32c6e542321423eeeb8eb6d265b428fc133d313b7baf7aee7 WatchSource:0}: Error finding container b9b0d534d481eba32c6e542321423eeeb8eb6d265b428fc133d313b7baf7aee7: Status 404 returned error can't find the container with id b9b0d534d481eba32c6e542321423eeeb8eb6d265b428fc133d313b7baf7aee7 Dec 03 13:24:40 crc kubenswrapper[4690]: I1203 13:24:40.159334 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn"] Dec 03 13:24:40 crc kubenswrapper[4690]: I1203 13:24:40.987699 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-85576f6988-gf7lj" event={"ID":"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b","Type":"ContainerStarted","Data":"2be21ce25ebf4636e316c1fa496b76d7009fcdad4ae5d6303bbe7d8ad226844c"} Dec 03 13:24:40 crc kubenswrapper[4690]: I1203 13:24:40.989404 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-85576f6988-gf7lj" event={"ID":"35bd2fae-21a1-4cf8-91fb-08a52c8d9e9b","Type":"ContainerStarted","Data":"1ea65eb092f1f5b0dde56794a877be449de87e1c689fbe438e15963ad4ee54a4"} Dec 03 13:24:40 crc kubenswrapper[4690]: I1203 13:24:40.989434 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" event={"ID":"33310d90-a18c-44db-a7f3-37f2086de63a","Type":"ContainerStarted","Data":"b9b0d534d481eba32c6e542321423eeeb8eb6d265b428fc133d313b7baf7aee7"} Dec 03 13:24:40 crc kubenswrapper[4690]: I1203 13:24:40.990763 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" event={"ID":"6020072e-2f65-4d13-85be-d300b4ec7f6d","Type":"ContainerStarted","Data":"15cb4d39c6b07b32941ad6159e2b80911be5f82ab234c592640c62f37c5b030e"} Dec 03 13:24:41 crc kubenswrapper[4690]: I1203 13:24:41.180698 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-85576f6988-gf7lj" podStartSLOduration=2.180680676 podStartE2EDuration="2.180680676s" podCreationTimestamp="2025-12-03 13:24:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:24:41.011999875 +0000 UTC m=+926.992920328" watchObservedRunningTime="2025-12-03 13:24:41.180680676 +0000 UTC m=+927.161601109" Dec 03 13:24:41 crc kubenswrapper[4690]: I1203 13:24:41.183835 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9hgpv"] Dec 03 13:24:41 crc kubenswrapper[4690]: I1203 13:24:41.184482 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9hgpv" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="registry-server" containerID="cri-o://839c5e04508699be9e21711763f4f80c9dc87155a8e4f74e1ff6c97539bc7a24" gracePeriod=2 Dec 03 13:24:41 crc kubenswrapper[4690]: I1203 13:24:41.998901 4690 generic.go:334] "Generic (PLEG): container finished" podID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerID="839c5e04508699be9e21711763f4f80c9dc87155a8e4f74e1ff6c97539bc7a24" exitCode=0 Dec 03 13:24:41 crc kubenswrapper[4690]: I1203 13:24:41.999138 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hgpv" event={"ID":"49d3e16a-82a0-4cc7-aadc-75496e4ab001","Type":"ContainerDied","Data":"839c5e04508699be9e21711763f4f80c9dc87155a8e4f74e1ff6c97539bc7a24"} Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.353780 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.511377 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-catalog-content\") pod \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.511949 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-utilities\") pod \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.512107 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xztz5\" (UniqueName: \"kubernetes.io/projected/49d3e16a-82a0-4cc7-aadc-75496e4ab001-kube-api-access-xztz5\") pod \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\" (UID: \"49d3e16a-82a0-4cc7-aadc-75496e4ab001\") " Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.513205 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-utilities" (OuterVolumeSpecName: "utilities") pod "49d3e16a-82a0-4cc7-aadc-75496e4ab001" (UID: "49d3e16a-82a0-4cc7-aadc-75496e4ab001"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.520465 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49d3e16a-82a0-4cc7-aadc-75496e4ab001-kube-api-access-xztz5" (OuterVolumeSpecName: "kube-api-access-xztz5") pod "49d3e16a-82a0-4cc7-aadc-75496e4ab001" (UID: "49d3e16a-82a0-4cc7-aadc-75496e4ab001"). InnerVolumeSpecName "kube-api-access-xztz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.615498 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.615537 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xztz5\" (UniqueName: \"kubernetes.io/projected/49d3e16a-82a0-4cc7-aadc-75496e4ab001-kube-api-access-xztz5\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.624682 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49d3e16a-82a0-4cc7-aadc-75496e4ab001" (UID: "49d3e16a-82a0-4cc7-aadc-75496e4ab001"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:24:42 crc kubenswrapper[4690]: I1203 13:24:42.716999 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d3e16a-82a0-4cc7-aadc-75496e4ab001-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:43 crc kubenswrapper[4690]: I1203 13:24:43.009556 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hgpv" event={"ID":"49d3e16a-82a0-4cc7-aadc-75496e4ab001","Type":"ContainerDied","Data":"283ffd4e36922c142c714bb62daab9984539f6117131e0bea5541f137f38b715"} Dec 03 13:24:43 crc kubenswrapper[4690]: I1203 13:24:43.009637 4690 scope.go:117] "RemoveContainer" containerID="839c5e04508699be9e21711763f4f80c9dc87155a8e4f74e1ff6c97539bc7a24" Dec 03 13:24:43 crc kubenswrapper[4690]: I1203 13:24:43.009704 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9hgpv" Dec 03 13:24:43 crc kubenswrapper[4690]: I1203 13:24:43.049493 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9hgpv"] Dec 03 13:24:43 crc kubenswrapper[4690]: I1203 13:24:43.058412 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9hgpv"] Dec 03 13:24:43 crc kubenswrapper[4690]: I1203 13:24:43.714538 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:43 crc kubenswrapper[4690]: I1203 13:24:43.714960 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:43 crc kubenswrapper[4690]: I1203 13:24:43.761994 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:44 crc kubenswrapper[4690]: I1203 13:24:44.056050 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:44 crc kubenswrapper[4690]: I1203 13:24:44.126397 4690 scope.go:117] "RemoveContainer" containerID="db9a40adb2499422f710936d9dd62c555817376427acf5b2e5100189079072f1" Dec 03 13:24:44 crc kubenswrapper[4690]: I1203 13:24:44.183651 4690 scope.go:117] "RemoveContainer" containerID="7e0cd575101f422989411e41ec37f80eb3d357a160ac14d82389f26c6a81216f" Dec 03 13:24:44 crc kubenswrapper[4690]: I1203 13:24:44.323849 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" path="/var/lib/kubelet/pods/49d3e16a-82a0-4cc7-aadc-75496e4ab001/volumes" Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.024751 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-v429n" event={"ID":"89d8054c-b35f-42fd-960d-16e1b39ad983","Type":"ContainerStarted","Data":"6931382db5ddd5fdbb4dd0e660350550be5940521db7f29f04a924d61b3192a6"} Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.025204 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.028747 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" event={"ID":"33310d90-a18c-44db-a7f3-37f2086de63a","Type":"ContainerStarted","Data":"79a608d28469c2ad55c249f725256ff7fc979c4a612637a178ab535c841feb91"} Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.030445 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" event={"ID":"bb3da90d-da37-49ad-ad46-230759602e46","Type":"ContainerStarted","Data":"962da6d57864ff8075c1c52a7f28dd94b2f8a07ea29f6c692a075188e0784a26"} Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.032455 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" event={"ID":"6020072e-2f65-4d13-85be-d300b4ec7f6d","Type":"ContainerStarted","Data":"23baba1a88b66406a88453905d916e4d0485bcea54c2590408607c97a6427cb4"} Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.044705 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-v429n" podStartSLOduration=2.223643103 podStartE2EDuration="7.044684184s" podCreationTimestamp="2025-12-03 13:24:38 +0000 UTC" firstStartedPulling="2025-12-03 13:24:39.381563861 +0000 UTC m=+925.362484294" lastFinishedPulling="2025-12-03 13:24:44.202604942 +0000 UTC m=+930.183525375" observedRunningTime="2025-12-03 13:24:45.042544359 +0000 UTC m=+931.023464802" watchObservedRunningTime="2025-12-03 13:24:45.044684184 +0000 UTC m=+931.025604617" Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.068548 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-4xg2d" podStartSLOduration=1.728633298 podStartE2EDuration="6.068523311s" podCreationTimestamp="2025-12-03 13:24:39 +0000 UTC" firstStartedPulling="2025-12-03 13:24:39.872423767 +0000 UTC m=+925.853344200" lastFinishedPulling="2025-12-03 13:24:44.21231379 +0000 UTC m=+930.193234213" observedRunningTime="2025-12-03 13:24:45.055785137 +0000 UTC m=+931.036705570" watchObservedRunningTime="2025-12-03 13:24:45.068523311 +0000 UTC m=+931.049443764" Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.784045 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" podStartSLOduration=3.742287646 podStartE2EDuration="7.784026686s" podCreationTimestamp="2025-12-03 13:24:38 +0000 UTC" firstStartedPulling="2025-12-03 13:24:40.173885624 +0000 UTC m=+926.154806057" lastFinishedPulling="2025-12-03 13:24:44.215624664 +0000 UTC m=+930.196545097" observedRunningTime="2025-12-03 13:24:45.072596825 +0000 UTC m=+931.053517258" watchObservedRunningTime="2025-12-03 13:24:45.784026686 +0000 UTC m=+931.764947119" Dec 03 13:24:45 crc kubenswrapper[4690]: I1203 13:24:45.784581 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4pmnk"] Dec 03 13:24:46 crc kubenswrapper[4690]: I1203 13:24:46.040139 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.048037 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" event={"ID":"33310d90-a18c-44db-a7f3-37f2086de63a","Type":"ContainerStarted","Data":"d554a2abb0cc07e0eda9055d8aabd9d2b9176cb05588a42f4fabddcb95ea653e"} Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.048310 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4pmnk" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerName="registry-server" containerID="cri-o://e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084" gracePeriod=2 Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.071218 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hjrl4" podStartSLOduration=2.391493663 podStartE2EDuration="9.071200128s" podCreationTimestamp="2025-12-03 13:24:38 +0000 UTC" firstStartedPulling="2025-12-03 13:24:39.999294032 +0000 UTC m=+925.980214465" lastFinishedPulling="2025-12-03 13:24:46.679000497 +0000 UTC m=+932.659920930" observedRunningTime="2025-12-03 13:24:47.067104304 +0000 UTC m=+933.048024757" watchObservedRunningTime="2025-12-03 13:24:47.071200128 +0000 UTC m=+933.052120551" Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.464593 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.594091 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d44sr\" (UniqueName: \"kubernetes.io/projected/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-kube-api-access-d44sr\") pod \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.594154 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-utilities\") pod \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.594357 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-catalog-content\") pod \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\" (UID: \"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e\") " Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.595568 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-utilities" (OuterVolumeSpecName: "utilities") pod "7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" (UID: "7e39f1af-1da1-44b5-a344-80bcdb5ddb8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.601082 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-kube-api-access-d44sr" (OuterVolumeSpecName: "kube-api-access-d44sr") pod "7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" (UID: "7e39f1af-1da1-44b5-a344-80bcdb5ddb8e"). InnerVolumeSpecName "kube-api-access-d44sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.647439 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" (UID: "7e39f1af-1da1-44b5-a344-80bcdb5ddb8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.697081 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d44sr\" (UniqueName: \"kubernetes.io/projected/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-kube-api-access-d44sr\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.697142 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:47 crc kubenswrapper[4690]: I1203 13:24:47.697163 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.058510 4690 generic.go:334] "Generic (PLEG): container finished" podID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerID="e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084" exitCode=0 Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.058561 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pmnk" event={"ID":"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e","Type":"ContainerDied","Data":"e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084"} Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.058613 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pmnk" event={"ID":"7e39f1af-1da1-44b5-a344-80bcdb5ddb8e","Type":"ContainerDied","Data":"93b311876337fb7918ef987baac8f8320e289e5282cebe7d7f0c5c56b9444fee"} Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.058632 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pmnk" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.058636 4690 scope.go:117] "RemoveContainer" containerID="e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.079652 4690 scope.go:117] "RemoveContainer" containerID="480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.098062 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4pmnk"] Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.103783 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4pmnk"] Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.113012 4690 scope.go:117] "RemoveContainer" containerID="f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.127775 4690 scope.go:117] "RemoveContainer" containerID="e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084" Dec 03 13:24:48 crc kubenswrapper[4690]: E1203 13:24:48.128379 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084\": container with ID starting with e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084 not found: ID does not exist" containerID="e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.128424 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084"} err="failed to get container status \"e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084\": rpc error: code = NotFound desc = could not find container \"e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084\": container with ID starting with e7cf3089b68501c474aa0811d940d82f7b586763244054f9c6a1f71eca937084 not found: ID does not exist" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.128456 4690 scope.go:117] "RemoveContainer" containerID="480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48" Dec 03 13:24:48 crc kubenswrapper[4690]: E1203 13:24:48.128715 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48\": container with ID starting with 480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48 not found: ID does not exist" containerID="480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.128743 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48"} err="failed to get container status \"480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48\": rpc error: code = NotFound desc = could not find container \"480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48\": container with ID starting with 480f33115d3e3b0c5ec8d2e4190fd4c254fb4535b6550d761c3cd9c36b9e1c48 not found: ID does not exist" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.128762 4690 scope.go:117] "RemoveContainer" containerID="f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869" Dec 03 13:24:48 crc kubenswrapper[4690]: E1203 13:24:48.129016 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869\": container with ID starting with f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869 not found: ID does not exist" containerID="f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.129043 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869"} err="failed to get container status \"f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869\": rpc error: code = NotFound desc = could not find container \"f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869\": container with ID starting with f84af07e9bcb31ea512494ae6cc6e8926d27e39f34794f02afb08ead189d3869 not found: ID does not exist" Dec 03 13:24:48 crc kubenswrapper[4690]: I1203 13:24:48.323759 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" path="/var/lib/kubelet/pods/7e39f1af-1da1-44b5-a344-80bcdb5ddb8e/volumes" Dec 03 13:24:49 crc kubenswrapper[4690]: I1203 13:24:49.368681 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-v429n" Dec 03 13:24:49 crc kubenswrapper[4690]: I1203 13:24:49.742816 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:49 crc kubenswrapper[4690]: I1203 13:24:49.742900 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:49 crc kubenswrapper[4690]: I1203 13:24:49.747575 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:50 crc kubenswrapper[4690]: I1203 13:24:50.079643 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-85576f6988-gf7lj" Dec 03 13:24:50 crc kubenswrapper[4690]: I1203 13:24:50.119630 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sqp4x"] Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.394789 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b4667"] Dec 03 13:24:52 crc kubenswrapper[4690]: E1203 13:24:52.395680 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="registry-server" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.395698 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="registry-server" Dec 03 13:24:52 crc kubenswrapper[4690]: E1203 13:24:52.395711 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="extract-content" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.395717 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="extract-content" Dec 03 13:24:52 crc kubenswrapper[4690]: E1203 13:24:52.395729 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerName="extract-utilities" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.395737 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerName="extract-utilities" Dec 03 13:24:52 crc kubenswrapper[4690]: E1203 13:24:52.395747 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerName="registry-server" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.395753 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerName="registry-server" Dec 03 13:24:52 crc kubenswrapper[4690]: E1203 13:24:52.395767 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="extract-utilities" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.395773 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="extract-utilities" Dec 03 13:24:52 crc kubenswrapper[4690]: E1203 13:24:52.395795 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerName="extract-content" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.395802 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerName="extract-content" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.395988 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e39f1af-1da1-44b5-a344-80bcdb5ddb8e" containerName="registry-server" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.396003 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d3e16a-82a0-4cc7-aadc-75496e4ab001" containerName="registry-server" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.397008 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.410745 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4667"] Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.582065 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q448g\" (UniqueName: \"kubernetes.io/projected/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-kube-api-access-q448g\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.582160 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-catalog-content\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.582198 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-utilities\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.683786 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q448g\" (UniqueName: \"kubernetes.io/projected/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-kube-api-access-q448g\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.683936 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-catalog-content\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.683989 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-utilities\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.684422 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-catalog-content\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.684550 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-utilities\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.710786 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q448g\" (UniqueName: \"kubernetes.io/projected/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-kube-api-access-q448g\") pod \"redhat-marketplace-b4667\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.715715 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:24:52 crc kubenswrapper[4690]: I1203 13:24:52.931381 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4667"] Dec 03 13:24:53 crc kubenswrapper[4690]: I1203 13:24:53.099044 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4667" event={"ID":"a84cf9b4-3630-4169-b23a-1aa681fd1e6b","Type":"ContainerStarted","Data":"9c65e214b76013a75fecbb09fe485d26fc5424792d6a6d01590b25955e306932"} Dec 03 13:24:54 crc kubenswrapper[4690]: I1203 13:24:54.107150 4690 generic.go:334] "Generic (PLEG): container finished" podID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerID="06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e" exitCode=0 Dec 03 13:24:54 crc kubenswrapper[4690]: I1203 13:24:54.107252 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4667" event={"ID":"a84cf9b4-3630-4169-b23a-1aa681fd1e6b","Type":"ContainerDied","Data":"06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e"} Dec 03 13:24:56 crc kubenswrapper[4690]: I1203 13:24:56.122746 4690 generic.go:334] "Generic (PLEG): container finished" podID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerID="4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36" exitCode=0 Dec 03 13:24:56 crc kubenswrapper[4690]: I1203 13:24:56.122784 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4667" event={"ID":"a84cf9b4-3630-4169-b23a-1aa681fd1e6b","Type":"ContainerDied","Data":"4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36"} Dec 03 13:24:57 crc kubenswrapper[4690]: I1203 13:24:57.136060 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4667" event={"ID":"a84cf9b4-3630-4169-b23a-1aa681fd1e6b","Type":"ContainerStarted","Data":"4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c"} Dec 03 13:24:59 crc kubenswrapper[4690]: I1203 13:24:59.927885 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-mw7tn" Dec 03 13:24:59 crc kubenswrapper[4690]: I1203 13:24:59.944326 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b4667" podStartSLOduration=5.465294075 podStartE2EDuration="7.944308327s" podCreationTimestamp="2025-12-03 13:24:52 +0000 UTC" firstStartedPulling="2025-12-03 13:24:54.108530752 +0000 UTC m=+940.089451185" lastFinishedPulling="2025-12-03 13:24:56.587545004 +0000 UTC m=+942.568465437" observedRunningTime="2025-12-03 13:24:57.158499263 +0000 UTC m=+943.139419706" watchObservedRunningTime="2025-12-03 13:24:59.944308327 +0000 UTC m=+945.925228760" Dec 03 13:25:02 crc kubenswrapper[4690]: I1203 13:25:02.716723 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:25:02 crc kubenswrapper[4690]: I1203 13:25:02.717161 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:25:02 crc kubenswrapper[4690]: I1203 13:25:02.761222 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:25:03 crc kubenswrapper[4690]: I1203 13:25:03.214440 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:25:05 crc kubenswrapper[4690]: I1203 13:25:05.585147 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4667"] Dec 03 13:25:05 crc kubenswrapper[4690]: I1203 13:25:05.585702 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b4667" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerName="registry-server" containerID="cri-o://4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c" gracePeriod=2 Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.086212 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.198727 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q448g\" (UniqueName: \"kubernetes.io/projected/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-kube-api-access-q448g\") pod \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.198813 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-catalog-content\") pod \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.198952 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-utilities\") pod \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\" (UID: \"a84cf9b4-3630-4169-b23a-1aa681fd1e6b\") " Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.200802 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-utilities" (OuterVolumeSpecName: "utilities") pod "a84cf9b4-3630-4169-b23a-1aa681fd1e6b" (UID: "a84cf9b4-3630-4169-b23a-1aa681fd1e6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.207185 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-kube-api-access-q448g" (OuterVolumeSpecName: "kube-api-access-q448g") pod "a84cf9b4-3630-4169-b23a-1aa681fd1e6b" (UID: "a84cf9b4-3630-4169-b23a-1aa681fd1e6b"). InnerVolumeSpecName "kube-api-access-q448g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.216144 4690 generic.go:334] "Generic (PLEG): container finished" podID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerID="4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c" exitCode=0 Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.216192 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4667" event={"ID":"a84cf9b4-3630-4169-b23a-1aa681fd1e6b","Type":"ContainerDied","Data":"4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c"} Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.216271 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b4667" event={"ID":"a84cf9b4-3630-4169-b23a-1aa681fd1e6b","Type":"ContainerDied","Data":"9c65e214b76013a75fecbb09fe485d26fc5424792d6a6d01590b25955e306932"} Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.216294 4690 scope.go:117] "RemoveContainer" containerID="4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.216429 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b4667" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.221802 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a84cf9b4-3630-4169-b23a-1aa681fd1e6b" (UID: "a84cf9b4-3630-4169-b23a-1aa681fd1e6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.240576 4690 scope.go:117] "RemoveContainer" containerID="4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.261687 4690 scope.go:117] "RemoveContainer" containerID="06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.302328 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.302375 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q448g\" (UniqueName: \"kubernetes.io/projected/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-kube-api-access-q448g\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.302396 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84cf9b4-3630-4169-b23a-1aa681fd1e6b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.304682 4690 scope.go:117] "RemoveContainer" containerID="4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c" Dec 03 13:25:07 crc kubenswrapper[4690]: E1203 13:25:07.305261 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c\": container with ID starting with 4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c not found: ID does not exist" containerID="4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.305287 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c"} err="failed to get container status \"4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c\": rpc error: code = NotFound desc = could not find container \"4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c\": container with ID starting with 4940ebeb8396a33790e4fe4b26f443e6e166014f81b5fa674d246f3f7a45265c not found: ID does not exist" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.305308 4690 scope.go:117] "RemoveContainer" containerID="4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36" Dec 03 13:25:07 crc kubenswrapper[4690]: E1203 13:25:07.305592 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36\": container with ID starting with 4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36 not found: ID does not exist" containerID="4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.305616 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36"} err="failed to get container status \"4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36\": rpc error: code = NotFound desc = could not find container \"4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36\": container with ID starting with 4e948246873600abaec3621ecaa10b57336fda9645d009b6db5e4cc610e55e36 not found: ID does not exist" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.305634 4690 scope.go:117] "RemoveContainer" containerID="06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e" Dec 03 13:25:07 crc kubenswrapper[4690]: E1203 13:25:07.306408 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e\": container with ID starting with 06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e not found: ID does not exist" containerID="06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.306439 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e"} err="failed to get container status \"06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e\": rpc error: code = NotFound desc = could not find container \"06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e\": container with ID starting with 06722583d50a86e498d9417b44c03b4b29e22e6683a67c224e1f7b21153b9e6e not found: ID does not exist" Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.546192 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4667"] Dec 03 13:25:07 crc kubenswrapper[4690]: I1203 13:25:07.553879 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b4667"] Dec 03 13:25:08 crc kubenswrapper[4690]: I1203 13:25:08.322933 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" path="/var/lib/kubelet/pods/a84cf9b4-3630-4169-b23a-1aa681fd1e6b/volumes" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.174089 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-sqp4x" podUID="fb8fb4ef-5880-4240-9b1a-d111f40d6e44" containerName="console" containerID="cri-o://bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5" gracePeriod=15 Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.693815 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sqp4x_fb8fb4ef-5880-4240-9b1a-d111f40d6e44/console/0.log" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.694163 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.813840 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-trusted-ca-bundle\") pod \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.813918 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-serving-cert\") pod \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.813987 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckfpz\" (UniqueName: \"kubernetes.io/projected/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-kube-api-access-ckfpz\") pod \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.814013 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-oauth-config\") pod \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.814094 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-config\") pod \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.814743 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-config" (OuterVolumeSpecName: "console-config") pod "fb8fb4ef-5880-4240-9b1a-d111f40d6e44" (UID: "fb8fb4ef-5880-4240-9b1a-d111f40d6e44"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.814752 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "fb8fb4ef-5880-4240-9b1a-d111f40d6e44" (UID: "fb8fb4ef-5880-4240-9b1a-d111f40d6e44"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.814811 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-oauth-serving-cert\") pod \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.814849 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-service-ca\") pod \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\" (UID: \"fb8fb4ef-5880-4240-9b1a-d111f40d6e44\") " Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.815265 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "fb8fb4ef-5880-4240-9b1a-d111f40d6e44" (UID: "fb8fb4ef-5880-4240-9b1a-d111f40d6e44"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.815471 4690 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.815492 4690 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.815503 4690 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.815717 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-service-ca" (OuterVolumeSpecName: "service-ca") pod "fb8fb4ef-5880-4240-9b1a-d111f40d6e44" (UID: "fb8fb4ef-5880-4240-9b1a-d111f40d6e44"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.820374 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-kube-api-access-ckfpz" (OuterVolumeSpecName: "kube-api-access-ckfpz") pod "fb8fb4ef-5880-4240-9b1a-d111f40d6e44" (UID: "fb8fb4ef-5880-4240-9b1a-d111f40d6e44"). InnerVolumeSpecName "kube-api-access-ckfpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.821402 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "fb8fb4ef-5880-4240-9b1a-d111f40d6e44" (UID: "fb8fb4ef-5880-4240-9b1a-d111f40d6e44"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.823286 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "fb8fb4ef-5880-4240-9b1a-d111f40d6e44" (UID: "fb8fb4ef-5880-4240-9b1a-d111f40d6e44"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.916811 4690 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.916847 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckfpz\" (UniqueName: \"kubernetes.io/projected/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-kube-api-access-ckfpz\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.916859 4690 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:15 crc kubenswrapper[4690]: I1203 13:25:15.916884 4690 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fb8fb4ef-5880-4240-9b1a-d111f40d6e44-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.279025 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sqp4x_fb8fb4ef-5880-4240-9b1a-d111f40d6e44/console/0.log" Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.279548 4690 generic.go:334] "Generic (PLEG): container finished" podID="fb8fb4ef-5880-4240-9b1a-d111f40d6e44" containerID="bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5" exitCode=2 Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.279606 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sqp4x" event={"ID":"fb8fb4ef-5880-4240-9b1a-d111f40d6e44","Type":"ContainerDied","Data":"bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5"} Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.279660 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sqp4x" event={"ID":"fb8fb4ef-5880-4240-9b1a-d111f40d6e44","Type":"ContainerDied","Data":"c2e5f4aeaade6389b6410accb890510d20d254b34b5623c9785ba36f0d2fda4b"} Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.279689 4690 scope.go:117] "RemoveContainer" containerID="bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5" Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.279707 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sqp4x" Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.307742 4690 scope.go:117] "RemoveContainer" containerID="bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5" Dec 03 13:25:16 crc kubenswrapper[4690]: E1203 13:25:16.310568 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5\": container with ID starting with bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5 not found: ID does not exist" containerID="bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5" Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.310801 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5"} err="failed to get container status \"bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5\": rpc error: code = NotFound desc = could not find container \"bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5\": container with ID starting with bdada800b4c635e08268135fa6ed4835b98f63d7bf80de6f5d6790730f316dc5 not found: ID does not exist" Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.327416 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sqp4x"] Dec 03 13:25:16 crc kubenswrapper[4690]: I1203 13:25:16.327946 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-sqp4x"] Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.836351 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng"] Dec 03 13:25:17 crc kubenswrapper[4690]: E1203 13:25:17.836587 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerName="extract-content" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.836600 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerName="extract-content" Dec 03 13:25:17 crc kubenswrapper[4690]: E1203 13:25:17.836610 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerName="extract-utilities" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.836616 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerName="extract-utilities" Dec 03 13:25:17 crc kubenswrapper[4690]: E1203 13:25:17.836633 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerName="registry-server" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.836638 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerName="registry-server" Dec 03 13:25:17 crc kubenswrapper[4690]: E1203 13:25:17.836651 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb8fb4ef-5880-4240-9b1a-d111f40d6e44" containerName="console" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.836657 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb8fb4ef-5880-4240-9b1a-d111f40d6e44" containerName="console" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.836762 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a84cf9b4-3630-4169-b23a-1aa681fd1e6b" containerName="registry-server" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.836779 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb8fb4ef-5880-4240-9b1a-d111f40d6e44" containerName="console" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.837615 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.841438 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.849998 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng"] Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.950467 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.950990 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:17 crc kubenswrapper[4690]: I1203 13:25:17.951027 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsbmg\" (UniqueName: \"kubernetes.io/projected/64127212-43bb-457c-b18d-3be4ea85d178-kube-api-access-xsbmg\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.052847 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.052906 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsbmg\" (UniqueName: \"kubernetes.io/projected/64127212-43bb-457c-b18d-3be4ea85d178-kube-api-access-xsbmg\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.052953 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.053431 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.053445 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.074029 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsbmg\" (UniqueName: \"kubernetes.io/projected/64127212-43bb-457c-b18d-3be4ea85d178-kube-api-access-xsbmg\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.160035 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.323632 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb8fb4ef-5880-4240-9b1a-d111f40d6e44" path="/var/lib/kubelet/pods/fb8fb4ef-5880-4240-9b1a-d111f40d6e44/volumes" Dec 03 13:25:18 crc kubenswrapper[4690]: I1203 13:25:18.621811 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng"] Dec 03 13:25:19 crc kubenswrapper[4690]: I1203 13:25:19.311480 4690 generic.go:334] "Generic (PLEG): container finished" podID="64127212-43bb-457c-b18d-3be4ea85d178" containerID="a103d445c8bf306c7bb444051485a115d4d8472066c36ee457ac0df37a57db29" exitCode=0 Dec 03 13:25:19 crc kubenswrapper[4690]: I1203 13:25:19.311524 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" event={"ID":"64127212-43bb-457c-b18d-3be4ea85d178","Type":"ContainerDied","Data":"a103d445c8bf306c7bb444051485a115d4d8472066c36ee457ac0df37a57db29"} Dec 03 13:25:19 crc kubenswrapper[4690]: I1203 13:25:19.311566 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" event={"ID":"64127212-43bb-457c-b18d-3be4ea85d178","Type":"ContainerStarted","Data":"d56ac64467325c1e77ce3bdb00d39e3f9242e748983fbbfc776998a5fc354eb1"} Dec 03 13:25:22 crc kubenswrapper[4690]: I1203 13:25:22.355477 4690 generic.go:334] "Generic (PLEG): container finished" podID="64127212-43bb-457c-b18d-3be4ea85d178" containerID="ac9d25ac5f44aa39a592e6d105bf0a442b9a7ef744eb76248d6ef7e8267a89d9" exitCode=0 Dec 03 13:25:22 crc kubenswrapper[4690]: I1203 13:25:22.356058 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" event={"ID":"64127212-43bb-457c-b18d-3be4ea85d178","Type":"ContainerDied","Data":"ac9d25ac5f44aa39a592e6d105bf0a442b9a7ef744eb76248d6ef7e8267a89d9"} Dec 03 13:25:23 crc kubenswrapper[4690]: I1203 13:25:23.367709 4690 generic.go:334] "Generic (PLEG): container finished" podID="64127212-43bb-457c-b18d-3be4ea85d178" containerID="5d01cbfa174801d69afdda91fa9109fca81a2b67eb4ac6ded18686b99e9018e9" exitCode=0 Dec 03 13:25:23 crc kubenswrapper[4690]: I1203 13:25:23.367788 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" event={"ID":"64127212-43bb-457c-b18d-3be4ea85d178","Type":"ContainerDied","Data":"5d01cbfa174801d69afdda91fa9109fca81a2b67eb4ac6ded18686b99e9018e9"} Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.775388 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.883098 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-util\") pod \"64127212-43bb-457c-b18d-3be4ea85d178\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.883177 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-bundle\") pod \"64127212-43bb-457c-b18d-3be4ea85d178\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.883222 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsbmg\" (UniqueName: \"kubernetes.io/projected/64127212-43bb-457c-b18d-3be4ea85d178-kube-api-access-xsbmg\") pod \"64127212-43bb-457c-b18d-3be4ea85d178\" (UID: \"64127212-43bb-457c-b18d-3be4ea85d178\") " Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.884375 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-bundle" (OuterVolumeSpecName: "bundle") pod "64127212-43bb-457c-b18d-3be4ea85d178" (UID: "64127212-43bb-457c-b18d-3be4ea85d178"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.889689 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64127212-43bb-457c-b18d-3be4ea85d178-kube-api-access-xsbmg" (OuterVolumeSpecName: "kube-api-access-xsbmg") pod "64127212-43bb-457c-b18d-3be4ea85d178" (UID: "64127212-43bb-457c-b18d-3be4ea85d178"). InnerVolumeSpecName "kube-api-access-xsbmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.897362 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-util" (OuterVolumeSpecName: "util") pod "64127212-43bb-457c-b18d-3be4ea85d178" (UID: "64127212-43bb-457c-b18d-3be4ea85d178"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.984728 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-util\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.985118 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64127212-43bb-457c-b18d-3be4ea85d178-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:24 crc kubenswrapper[4690]: I1203 13:25:24.985130 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsbmg\" (UniqueName: \"kubernetes.io/projected/64127212-43bb-457c-b18d-3be4ea85d178-kube-api-access-xsbmg\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:25 crc kubenswrapper[4690]: I1203 13:25:25.383319 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" event={"ID":"64127212-43bb-457c-b18d-3be4ea85d178","Type":"ContainerDied","Data":"d56ac64467325c1e77ce3bdb00d39e3f9242e748983fbbfc776998a5fc354eb1"} Dec 03 13:25:25 crc kubenswrapper[4690]: I1203 13:25:25.383365 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d56ac64467325c1e77ce3bdb00d39e3f9242e748983fbbfc776998a5fc354eb1" Dec 03 13:25:25 crc kubenswrapper[4690]: I1203 13:25:25.383457 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.146308 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6"] Dec 03 13:25:34 crc kubenswrapper[4690]: E1203 13:25:34.147324 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64127212-43bb-457c-b18d-3be4ea85d178" containerName="extract" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.147343 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="64127212-43bb-457c-b18d-3be4ea85d178" containerName="extract" Dec 03 13:25:34 crc kubenswrapper[4690]: E1203 13:25:34.147359 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64127212-43bb-457c-b18d-3be4ea85d178" containerName="util" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.147367 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="64127212-43bb-457c-b18d-3be4ea85d178" containerName="util" Dec 03 13:25:34 crc kubenswrapper[4690]: E1203 13:25:34.147384 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64127212-43bb-457c-b18d-3be4ea85d178" containerName="pull" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.147393 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="64127212-43bb-457c-b18d-3be4ea85d178" containerName="pull" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.147541 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="64127212-43bb-457c-b18d-3be4ea85d178" containerName="extract" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.148185 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.150806 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-67r5f" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.151042 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.151752 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.152307 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.159005 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.214299 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6"] Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.226890 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b62fccc-0740-4e1f-8efb-61bac3a9261e-webhook-cert\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.227021 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxfsj\" (UniqueName: \"kubernetes.io/projected/6b62fccc-0740-4e1f-8efb-61bac3a9261e-kube-api-access-wxfsj\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.227099 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b62fccc-0740-4e1f-8efb-61bac3a9261e-apiservice-cert\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.325358 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m"] Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.326237 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.328457 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxfsj\" (UniqueName: \"kubernetes.io/projected/6b62fccc-0740-4e1f-8efb-61bac3a9261e-kube-api-access-wxfsj\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.328531 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b62fccc-0740-4e1f-8efb-61bac3a9261e-apiservice-cert\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.328580 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b62fccc-0740-4e1f-8efb-61bac3a9261e-webhook-cert\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.330223 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.330787 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-4brvb" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.330975 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.338777 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b62fccc-0740-4e1f-8efb-61bac3a9261e-apiservice-cert\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.344659 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m"] Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.353781 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxfsj\" (UniqueName: \"kubernetes.io/projected/6b62fccc-0740-4e1f-8efb-61bac3a9261e-kube-api-access-wxfsj\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.354675 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b62fccc-0740-4e1f-8efb-61bac3a9261e-webhook-cert\") pod \"metallb-operator-controller-manager-5994f497d6-5c9m6\" (UID: \"6b62fccc-0740-4e1f-8efb-61bac3a9261e\") " pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.430634 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93d03eb3-9a1d-452e-8812-c5a871652ab3-apiservice-cert\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.430702 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93d03eb3-9a1d-452e-8812-c5a871652ab3-webhook-cert\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.430743 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxmq2\" (UniqueName: \"kubernetes.io/projected/93d03eb3-9a1d-452e-8812-c5a871652ab3-kube-api-access-rxmq2\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.469445 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.532888 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxmq2\" (UniqueName: \"kubernetes.io/projected/93d03eb3-9a1d-452e-8812-c5a871652ab3-kube-api-access-rxmq2\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.533064 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93d03eb3-9a1d-452e-8812-c5a871652ab3-apiservice-cert\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.533098 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93d03eb3-9a1d-452e-8812-c5a871652ab3-webhook-cert\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.542950 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93d03eb3-9a1d-452e-8812-c5a871652ab3-apiservice-cert\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.543039 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93d03eb3-9a1d-452e-8812-c5a871652ab3-webhook-cert\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.555269 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxmq2\" (UniqueName: \"kubernetes.io/projected/93d03eb3-9a1d-452e-8812-c5a871652ab3-kube-api-access-rxmq2\") pod \"metallb-operator-webhook-server-c954657ff-b5h2m\" (UID: \"93d03eb3-9a1d-452e-8812-c5a871652ab3\") " pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.702807 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:34 crc kubenswrapper[4690]: I1203 13:25:34.907855 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6"] Dec 03 13:25:34 crc kubenswrapper[4690]: W1203 13:25:34.918477 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b62fccc_0740_4e1f_8efb_61bac3a9261e.slice/crio-0bf2c11e1f55584c6f5aa860e064b28e11e2f782bcb132d87e4bd0ab18dba80a WatchSource:0}: Error finding container 0bf2c11e1f55584c6f5aa860e064b28e11e2f782bcb132d87e4bd0ab18dba80a: Status 404 returned error can't find the container with id 0bf2c11e1f55584c6f5aa860e064b28e11e2f782bcb132d87e4bd0ab18dba80a Dec 03 13:25:35 crc kubenswrapper[4690]: I1203 13:25:35.154939 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m"] Dec 03 13:25:35 crc kubenswrapper[4690]: W1203 13:25:35.156137 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93d03eb3_9a1d_452e_8812_c5a871652ab3.slice/crio-c08b8817514cb70aefc8d37e658be8e280e4f978d19fe1873864f073a6c078ad WatchSource:0}: Error finding container c08b8817514cb70aefc8d37e658be8e280e4f978d19fe1873864f073a6c078ad: Status 404 returned error can't find the container with id c08b8817514cb70aefc8d37e658be8e280e4f978d19fe1873864f073a6c078ad Dec 03 13:25:35 crc kubenswrapper[4690]: I1203 13:25:35.446329 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" event={"ID":"93d03eb3-9a1d-452e-8812-c5a871652ab3","Type":"ContainerStarted","Data":"c08b8817514cb70aefc8d37e658be8e280e4f978d19fe1873864f073a6c078ad"} Dec 03 13:25:35 crc kubenswrapper[4690]: I1203 13:25:35.448192 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" event={"ID":"6b62fccc-0740-4e1f-8efb-61bac3a9261e","Type":"ContainerStarted","Data":"0bf2c11e1f55584c6f5aa860e064b28e11e2f782bcb132d87e4bd0ab18dba80a"} Dec 03 13:25:40 crc kubenswrapper[4690]: I1203 13:25:40.495919 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" event={"ID":"6b62fccc-0740-4e1f-8efb-61bac3a9261e","Type":"ContainerStarted","Data":"09f8969553d0ed008090fed23d6851950c9af3ded3f129f01c0ec533374a8177"} Dec 03 13:25:40 crc kubenswrapper[4690]: I1203 13:25:40.497606 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:25:40 crc kubenswrapper[4690]: I1203 13:25:40.497823 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" event={"ID":"93d03eb3-9a1d-452e-8812-c5a871652ab3","Type":"ContainerStarted","Data":"a80afd32abeebf672ab6df16d5f5fc2d8ce3ea4da01bd814913eacb1df10f939"} Dec 03 13:25:40 crc kubenswrapper[4690]: I1203 13:25:40.498012 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:25:40 crc kubenswrapper[4690]: I1203 13:25:40.523333 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" podStartSLOduration=1.2900425819999999 podStartE2EDuration="6.523309725s" podCreationTimestamp="2025-12-03 13:25:34 +0000 UTC" firstStartedPulling="2025-12-03 13:25:34.923107671 +0000 UTC m=+980.904028104" lastFinishedPulling="2025-12-03 13:25:40.156374814 +0000 UTC m=+986.137295247" observedRunningTime="2025-12-03 13:25:40.518334098 +0000 UTC m=+986.499254541" watchObservedRunningTime="2025-12-03 13:25:40.523309725 +0000 UTC m=+986.504230158" Dec 03 13:25:40 crc kubenswrapper[4690]: I1203 13:25:40.544078 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" podStartSLOduration=1.529873882 podStartE2EDuration="6.544055773s" podCreationTimestamp="2025-12-03 13:25:34 +0000 UTC" firstStartedPulling="2025-12-03 13:25:35.159211937 +0000 UTC m=+981.140132360" lastFinishedPulling="2025-12-03 13:25:40.173393818 +0000 UTC m=+986.154314251" observedRunningTime="2025-12-03 13:25:40.536738957 +0000 UTC m=+986.517659390" watchObservedRunningTime="2025-12-03 13:25:40.544055773 +0000 UTC m=+986.524976206" Dec 03 13:25:46 crc kubenswrapper[4690]: I1203 13:25:46.823486 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:25:46 crc kubenswrapper[4690]: I1203 13:25:46.824107 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:25:54 crc kubenswrapper[4690]: I1203 13:25:54.708180 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-c954657ff-b5h2m" Dec 03 13:26:14 crc kubenswrapper[4690]: I1203 13:26:14.473924 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5994f497d6-5c9m6" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.237362 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hfmdl"] Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.240914 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.242499 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv"] Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.243360 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.251425 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.254250 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.254503 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-cf4p8" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.257746 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.262979 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv"] Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279392 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-startup\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279498 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcbwn\" (UniqueName: \"kubernetes.io/projected/ece18280-b324-4c1e-b3d5-411e5c146869-kube-api-access-lcbwn\") pod \"frr-k8s-webhook-server-7fcb986d4-r8zcv\" (UID: \"ece18280-b324-4c1e-b3d5-411e5c146869\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279543 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqs99\" (UniqueName: \"kubernetes.io/projected/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-kube-api-access-rqs99\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279573 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279605 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-conf\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279629 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-sockets\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279679 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ece18280-b324-4c1e-b3d5-411e5c146869-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-r8zcv\" (UID: \"ece18280-b324-4c1e-b3d5-411e5c146869\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279735 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-reloader\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.279757 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.377938 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-7hmjp"] Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.379300 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.380966 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcbwn\" (UniqueName: \"kubernetes.io/projected/ece18280-b324-4c1e-b3d5-411e5c146869-kube-api-access-lcbwn\") pod \"frr-k8s-webhook-server-7fcb986d4-r8zcv\" (UID: \"ece18280-b324-4c1e-b3d5-411e5c146869\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381014 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqs99\" (UniqueName: \"kubernetes.io/projected/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-kube-api-access-rqs99\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381038 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381065 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-conf\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381082 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-sockets\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381107 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ece18280-b324-4c1e-b3d5-411e5c146869-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-r8zcv\" (UID: \"ece18280-b324-4c1e-b3d5-411e5c146869\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381130 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381157 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-reloader\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381175 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381196 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1b3ec2d-089c-409a-901e-4f466ec07855-metallb-excludel2\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381213 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-metrics-certs\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381234 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-startup\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.381251 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcbd2\" (UniqueName: \"kubernetes.io/projected/e1b3ec2d-089c-409a-901e-4f466ec07855-kube-api-access-gcbd2\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.382177 4690 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.382222 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs podName:efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2 nodeName:}" failed. No retries permitted until 2025-12-03 13:26:15.882207387 +0000 UTC m=+1021.863127820 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs") pod "frr-k8s-hfmdl" (UID: "efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2") : secret "frr-k8s-certs-secret" not found Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.382642 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-conf\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.382822 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-sockets\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.382891 4690 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.386633 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-reloader\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.387122 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.388484 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-frr-startup\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.389220 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ece18280-b324-4c1e-b3d5-411e5c146869-cert podName:ece18280-b324-4c1e-b3d5-411e5c146869 nodeName:}" failed. No retries permitted until 2025-12-03 13:26:15.889199505 +0000 UTC m=+1021.870119938 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ece18280-b324-4c1e-b3d5-411e5c146869-cert") pod "frr-k8s-webhook-server-7fcb986d4-r8zcv" (UID: "ece18280-b324-4c1e-b3d5-411e5c146869") : secret "frr-k8s-webhook-server-cert" not found Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.389755 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.390201 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.391971 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.392302 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-9wxdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.409471 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-t4xcd"] Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.415282 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.420034 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.420858 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqs99\" (UniqueName: \"kubernetes.io/projected/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-kube-api-access-rqs99\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.430721 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcbwn\" (UniqueName: \"kubernetes.io/projected/ece18280-b324-4c1e-b3d5-411e5c146869-kube-api-access-lcbwn\") pod \"frr-k8s-webhook-server-7fcb986d4-r8zcv\" (UID: \"ece18280-b324-4c1e-b3d5-411e5c146869\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.453620 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-t4xcd"] Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.482172 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcbd2\" (UniqueName: \"kubernetes.io/projected/e1b3ec2d-089c-409a-901e-4f466ec07855-kube-api-access-gcbd2\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.482294 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-cert\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.482324 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-metrics-certs\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.482359 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.482386 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbxwl\" (UniqueName: \"kubernetes.io/projected/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-kube-api-access-kbxwl\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.482432 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1b3ec2d-089c-409a-901e-4f466ec07855-metallb-excludel2\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.482456 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-metrics-certs\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.482592 4690 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.482655 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-metrics-certs podName:e1b3ec2d-089c-409a-901e-4f466ec07855 nodeName:}" failed. No retries permitted until 2025-12-03 13:26:15.982633396 +0000 UTC m=+1021.963553829 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-metrics-certs") pod "speaker-7hmjp" (UID: "e1b3ec2d-089c-409a-901e-4f466ec07855") : secret "speaker-certs-secret" not found Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.483098 4690 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.483210 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist podName:e1b3ec2d-089c-409a-901e-4f466ec07855 nodeName:}" failed. No retries permitted until 2025-12-03 13:26:15.98317958 +0000 UTC m=+1021.964100193 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist") pod "speaker-7hmjp" (UID: "e1b3ec2d-089c-409a-901e-4f466ec07855") : secret "metallb-memberlist" not found Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.483644 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1b3ec2d-089c-409a-901e-4f466ec07855-metallb-excludel2\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.532460 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcbd2\" (UniqueName: \"kubernetes.io/projected/e1b3ec2d-089c-409a-901e-4f466ec07855-kube-api-access-gcbd2\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.583885 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-cert\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.583957 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-metrics-certs\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.584026 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbxwl\" (UniqueName: \"kubernetes.io/projected/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-kube-api-access-kbxwl\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.585807 4690 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.585898 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-metrics-certs podName:e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1 nodeName:}" failed. No retries permitted until 2025-12-03 13:26:16.085857897 +0000 UTC m=+1022.066778340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-metrics-certs") pod "controller-f8648f98b-t4xcd" (UID: "e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1") : secret "controller-certs-secret" not found Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.589314 4690 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.608999 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbxwl\" (UniqueName: \"kubernetes.io/projected/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-kube-api-access-kbxwl\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.609363 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-cert\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.889442 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.889524 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ece18280-b324-4c1e-b3d5-411e5c146869-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-r8zcv\" (UID: \"ece18280-b324-4c1e-b3d5-411e5c146869\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.889654 4690 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.889752 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs podName:efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2 nodeName:}" failed. No retries permitted until 2025-12-03 13:26:16.88972877 +0000 UTC m=+1022.870649373 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs") pod "frr-k8s-hfmdl" (UID: "efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2") : secret "frr-k8s-certs-secret" not found Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.894467 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ece18280-b324-4c1e-b3d5-411e5c146869-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-r8zcv\" (UID: \"ece18280-b324-4c1e-b3d5-411e5c146869\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.990555 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-metrics-certs\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.990713 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.990897 4690 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 13:26:15 crc kubenswrapper[4690]: E1203 13:26:15.990975 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist podName:e1b3ec2d-089c-409a-901e-4f466ec07855 nodeName:}" failed. No retries permitted until 2025-12-03 13:26:16.99095519 +0000 UTC m=+1022.971875623 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist") pod "speaker-7hmjp" (UID: "e1b3ec2d-089c-409a-901e-4f466ec07855") : secret "metallb-memberlist" not found Dec 03 13:26:15 crc kubenswrapper[4690]: I1203 13:26:15.994183 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-metrics-certs\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.092055 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-metrics-certs\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.095626 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1-metrics-certs\") pod \"controller-f8648f98b-t4xcd\" (UID: \"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1\") " pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.104069 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.165463 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.547312 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-t4xcd"] Dec 03 13:26:16 crc kubenswrapper[4690]: W1203 13:26:16.550207 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode65a2f57_73b0_4eb7_a62e_c8f4b8f816a1.slice/crio-9a7c6d37d4e778b2c36c7f7a9bfe607f571f4f6d029ca36c44919d23c9b69a45 WatchSource:0}: Error finding container 9a7c6d37d4e778b2c36c7f7a9bfe607f571f4f6d029ca36c44919d23c9b69a45: Status 404 returned error can't find the container with id 9a7c6d37d4e778b2c36c7f7a9bfe607f571f4f6d029ca36c44919d23c9b69a45 Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.643197 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv"] Dec 03 13:26:16 crc kubenswrapper[4690]: W1203 13:26:16.654723 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podece18280_b324_4c1e_b3d5_411e5c146869.slice/crio-1232ff09ab013e7fc8a6dcf5c57d638bbc0579df192e1219b5989957e4ad69ce WatchSource:0}: Error finding container 1232ff09ab013e7fc8a6dcf5c57d638bbc0579df192e1219b5989957e4ad69ce: Status 404 returned error can't find the container with id 1232ff09ab013e7fc8a6dcf5c57d638bbc0579df192e1219b5989957e4ad69ce Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.657465 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.749627 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t4xcd" event={"ID":"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1","Type":"ContainerStarted","Data":"43346a76599aeb3abdd1d59612c24cb5c13aef3a8138522bb53b13663b8c675a"} Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.749687 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t4xcd" event={"ID":"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1","Type":"ContainerStarted","Data":"9a7c6d37d4e778b2c36c7f7a9bfe607f571f4f6d029ca36c44919d23c9b69a45"} Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.751129 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" event={"ID":"ece18280-b324-4c1e-b3d5-411e5c146869","Type":"ContainerStarted","Data":"1232ff09ab013e7fc8a6dcf5c57d638bbc0579df192e1219b5989957e4ad69ce"} Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.823666 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.823742 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.908687 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:16 crc kubenswrapper[4690]: I1203 13:26:16.916907 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2-metrics-certs\") pod \"frr-k8s-hfmdl\" (UID: \"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2\") " pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.010591 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.014307 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1b3ec2d-089c-409a-901e-4f466ec07855-memberlist\") pod \"speaker-7hmjp\" (UID: \"e1b3ec2d-089c-409a-901e-4f466ec07855\") " pod="metallb-system/speaker-7hmjp" Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.065074 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.200664 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7hmjp" Dec 03 13:26:17 crc kubenswrapper[4690]: W1203 13:26:17.224264 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1b3ec2d_089c_409a_901e_4f466ec07855.slice/crio-2e4544a65a887d8ff6aa1cf72d64da3e2cbb3f4ee5dcf209f7822a397ba53f21 WatchSource:0}: Error finding container 2e4544a65a887d8ff6aa1cf72d64da3e2cbb3f4ee5dcf209f7822a397ba53f21: Status 404 returned error can't find the container with id 2e4544a65a887d8ff6aa1cf72d64da3e2cbb3f4ee5dcf209f7822a397ba53f21 Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.770010 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t4xcd" event={"ID":"e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1","Type":"ContainerStarted","Data":"4776eb26868e7c8a40468f64826eb83e321a8bf2eafc0267c9e00ceb9848c404"} Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.772188 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.784440 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7hmjp" event={"ID":"e1b3ec2d-089c-409a-901e-4f466ec07855","Type":"ContainerStarted","Data":"2b3eae400618866261ed40a8d1164b0be780806bb98b886e97a28d4c81e39f38"} Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.784502 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7hmjp" event={"ID":"e1b3ec2d-089c-409a-901e-4f466ec07855","Type":"ContainerStarted","Data":"2e4544a65a887d8ff6aa1cf72d64da3e2cbb3f4ee5dcf209f7822a397ba53f21"} Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.788260 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerStarted","Data":"485f47cec637f9cc83f6d277b98684d0af9f2875ee7cc11f4e1456da18cdb0f0"} Dec 03 13:26:17 crc kubenswrapper[4690]: I1203 13:26:17.804284 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-t4xcd" podStartSLOduration=2.804266007 podStartE2EDuration="2.804266007s" podCreationTimestamp="2025-12-03 13:26:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:26:17.802167053 +0000 UTC m=+1023.783087486" watchObservedRunningTime="2025-12-03 13:26:17.804266007 +0000 UTC m=+1023.785186440" Dec 03 13:26:18 crc kubenswrapper[4690]: I1203 13:26:18.798669 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7hmjp" event={"ID":"e1b3ec2d-089c-409a-901e-4f466ec07855","Type":"ContainerStarted","Data":"6c6b7b5bf9b620c46941499a6deb7d69e9771bcfecbb22a2c5422cd7855d198a"} Dec 03 13:26:18 crc kubenswrapper[4690]: I1203 13:26:18.798947 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-7hmjp" Dec 03 13:26:18 crc kubenswrapper[4690]: I1203 13:26:18.821183 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-7hmjp" podStartSLOduration=3.821155718 podStartE2EDuration="3.821155718s" podCreationTimestamp="2025-12-03 13:26:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:26:18.816234573 +0000 UTC m=+1024.797155006" watchObservedRunningTime="2025-12-03 13:26:18.821155718 +0000 UTC m=+1024.802076151" Dec 03 13:26:24 crc kubenswrapper[4690]: I1203 13:26:24.848116 4690 generic.go:334] "Generic (PLEG): container finished" podID="efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2" containerID="c630351f39d48d4cf019aac4a904b2aea9dc6ebac07fa3ec22331d1ba44deb17" exitCode=0 Dec 03 13:26:24 crc kubenswrapper[4690]: I1203 13:26:24.848263 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerDied","Data":"c630351f39d48d4cf019aac4a904b2aea9dc6ebac07fa3ec22331d1ba44deb17"} Dec 03 13:26:24 crc kubenswrapper[4690]: I1203 13:26:24.850690 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" event={"ID":"ece18280-b324-4c1e-b3d5-411e5c146869","Type":"ContainerStarted","Data":"02cfb441c300b5b0d648d99aa3fca0bc650aa34c671d858ed3dafa0d6eb6cf76"} Dec 03 13:26:24 crc kubenswrapper[4690]: I1203 13:26:24.850838 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:24 crc kubenswrapper[4690]: I1203 13:26:24.934704 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" podStartSLOduration=2.881108095 podStartE2EDuration="9.934684223s" podCreationTimestamp="2025-12-03 13:26:15 +0000 UTC" firstStartedPulling="2025-12-03 13:26:16.657170476 +0000 UTC m=+1022.638090909" lastFinishedPulling="2025-12-03 13:26:23.710746604 +0000 UTC m=+1029.691667037" observedRunningTime="2025-12-03 13:26:24.928456834 +0000 UTC m=+1030.909377267" watchObservedRunningTime="2025-12-03 13:26:24.934684223 +0000 UTC m=+1030.915604656" Dec 03 13:26:25 crc kubenswrapper[4690]: I1203 13:26:25.858511 4690 generic.go:334] "Generic (PLEG): container finished" podID="efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2" containerID="3ff83af42561a17f7ef96105dad3eceed804a065bf84c025d02d0f979310caff" exitCode=0 Dec 03 13:26:25 crc kubenswrapper[4690]: I1203 13:26:25.858603 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerDied","Data":"3ff83af42561a17f7ef96105dad3eceed804a065bf84c025d02d0f979310caff"} Dec 03 13:26:26 crc kubenswrapper[4690]: I1203 13:26:26.110501 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-t4xcd" Dec 03 13:26:26 crc kubenswrapper[4690]: I1203 13:26:26.867158 4690 generic.go:334] "Generic (PLEG): container finished" podID="efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2" containerID="f90900aa2c8b196e3c46d7510dd1ab7780bbf73ba93b0c7b2873a5b5b7b1d938" exitCode=0 Dec 03 13:26:26 crc kubenswrapper[4690]: I1203 13:26:26.867369 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerDied","Data":"f90900aa2c8b196e3c46d7510dd1ab7780bbf73ba93b0c7b2873a5b5b7b1d938"} Dec 03 13:26:27 crc kubenswrapper[4690]: I1203 13:26:27.204587 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-7hmjp" Dec 03 13:26:27 crc kubenswrapper[4690]: I1203 13:26:27.881989 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerStarted","Data":"4d5630f553e48f61fd0d8ff9d3befb0f88627924b5de20cf7b7c11ac0a914e5e"} Dec 03 13:26:27 crc kubenswrapper[4690]: I1203 13:26:27.882826 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerStarted","Data":"1bb857138404367a08661cc35201bc9d5150717a10ddbb6296091dc485385dc1"} Dec 03 13:26:27 crc kubenswrapper[4690]: I1203 13:26:27.882843 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerStarted","Data":"172d4b12f3d2b99bced3d54164ee8aa1f5f40c0ac5cf559dbca512633a5ac344"} Dec 03 13:26:27 crc kubenswrapper[4690]: I1203 13:26:27.882891 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerStarted","Data":"1e268906a6e694bbe9e377a0fc8f08e8d20908ef7e184a99804cb39850d7b9c3"} Dec 03 13:26:27 crc kubenswrapper[4690]: I1203 13:26:27.882903 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerStarted","Data":"8d9f12fb431cc5d4a6d261f3acbb10cbf0d915f7d9c7de29ce72283d36d29d28"} Dec 03 13:26:28 crc kubenswrapper[4690]: I1203 13:26:28.892616 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hfmdl" event={"ID":"efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2","Type":"ContainerStarted","Data":"ef11999a08a69249fb2c36c8df006abdd95716477e04256ce12781e56a5f161a"} Dec 03 13:26:28 crc kubenswrapper[4690]: I1203 13:26:28.893163 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:28 crc kubenswrapper[4690]: I1203 13:26:28.919599 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hfmdl" podStartSLOduration=7.3700894120000005 podStartE2EDuration="13.919579315s" podCreationTimestamp="2025-12-03 13:26:15 +0000 UTC" firstStartedPulling="2025-12-03 13:26:17.182114403 +0000 UTC m=+1023.163034836" lastFinishedPulling="2025-12-03 13:26:23.731604316 +0000 UTC m=+1029.712524739" observedRunningTime="2025-12-03 13:26:28.915926392 +0000 UTC m=+1034.896846845" watchObservedRunningTime="2025-12-03 13:26:28.919579315 +0000 UTC m=+1034.900499748" Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.650762 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pl6hw"] Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.652057 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pl6hw" Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.654140 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.660371 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.660469 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-qzbcx" Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.709889 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pl6hw"] Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.752975 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v7d6\" (UniqueName: \"kubernetes.io/projected/a58fc074-d792-4cec-87c7-b13197c39733-kube-api-access-4v7d6\") pod \"openstack-operator-index-pl6hw\" (UID: \"a58fc074-d792-4cec-87c7-b13197c39733\") " pod="openstack-operators/openstack-operator-index-pl6hw" Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.854219 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v7d6\" (UniqueName: \"kubernetes.io/projected/a58fc074-d792-4cec-87c7-b13197c39733-kube-api-access-4v7d6\") pod \"openstack-operator-index-pl6hw\" (UID: \"a58fc074-d792-4cec-87c7-b13197c39733\") " pod="openstack-operators/openstack-operator-index-pl6hw" Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.879431 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v7d6\" (UniqueName: \"kubernetes.io/projected/a58fc074-d792-4cec-87c7-b13197c39733-kube-api-access-4v7d6\") pod \"openstack-operator-index-pl6hw\" (UID: \"a58fc074-d792-4cec-87c7-b13197c39733\") " pod="openstack-operators/openstack-operator-index-pl6hw" Dec 03 13:26:30 crc kubenswrapper[4690]: I1203 13:26:30.978942 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pl6hw" Dec 03 13:26:31 crc kubenswrapper[4690]: I1203 13:26:31.723852 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pl6hw"] Dec 03 13:26:31 crc kubenswrapper[4690]: I1203 13:26:31.913376 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pl6hw" event={"ID":"a58fc074-d792-4cec-87c7-b13197c39733","Type":"ContainerStarted","Data":"2906729a55d65584c0b2c34c0a528eff60b8d3b632fe56f70ba997306abca1b4"} Dec 03 13:26:32 crc kubenswrapper[4690]: I1203 13:26:32.067725 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:32 crc kubenswrapper[4690]: I1203 13:26:32.114949 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:33 crc kubenswrapper[4690]: I1203 13:26:33.814322 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-pl6hw"] Dec 03 13:26:34 crc kubenswrapper[4690]: I1203 13:26:34.434946 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-95zhf"] Dec 03 13:26:34 crc kubenswrapper[4690]: I1203 13:26:34.436585 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:34 crc kubenswrapper[4690]: I1203 13:26:34.438861 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-95zhf"] Dec 03 13:26:34 crc kubenswrapper[4690]: I1203 13:26:34.618404 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9lj4\" (UniqueName: \"kubernetes.io/projected/ff4c17ee-a756-4743-aeb8-aae8fe520712-kube-api-access-b9lj4\") pod \"openstack-operator-index-95zhf\" (UID: \"ff4c17ee-a756-4743-aeb8-aae8fe520712\") " pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:34 crc kubenswrapper[4690]: I1203 13:26:34.719918 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9lj4\" (UniqueName: \"kubernetes.io/projected/ff4c17ee-a756-4743-aeb8-aae8fe520712-kube-api-access-b9lj4\") pod \"openstack-operator-index-95zhf\" (UID: \"ff4c17ee-a756-4743-aeb8-aae8fe520712\") " pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:34 crc kubenswrapper[4690]: I1203 13:26:34.747273 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9lj4\" (UniqueName: \"kubernetes.io/projected/ff4c17ee-a756-4743-aeb8-aae8fe520712-kube-api-access-b9lj4\") pod \"openstack-operator-index-95zhf\" (UID: \"ff4c17ee-a756-4743-aeb8-aae8fe520712\") " pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:34 crc kubenswrapper[4690]: I1203 13:26:34.780915 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:35 crc kubenswrapper[4690]: I1203 13:26:35.217640 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-95zhf"] Dec 03 13:26:35 crc kubenswrapper[4690]: W1203 13:26:35.222219 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff4c17ee_a756_4743_aeb8_aae8fe520712.slice/crio-6c9c9e52d254001e15d2457da474aa653fa989c9f313b7a7aad4cf83bd7906e9 WatchSource:0}: Error finding container 6c9c9e52d254001e15d2457da474aa653fa989c9f313b7a7aad4cf83bd7906e9: Status 404 returned error can't find the container with id 6c9c9e52d254001e15d2457da474aa653fa989c9f313b7a7aad4cf83bd7906e9 Dec 03 13:26:35 crc kubenswrapper[4690]: I1203 13:26:35.946091 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pl6hw" event={"ID":"a58fc074-d792-4cec-87c7-b13197c39733","Type":"ContainerStarted","Data":"49dc2db42f852366b40cc26f12ebb86b15d30c3bcc8fc0624bd9b38b48914fbd"} Dec 03 13:26:35 crc kubenswrapper[4690]: I1203 13:26:35.946290 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-pl6hw" podUID="a58fc074-d792-4cec-87c7-b13197c39733" containerName="registry-server" containerID="cri-o://49dc2db42f852366b40cc26f12ebb86b15d30c3bcc8fc0624bd9b38b48914fbd" gracePeriod=2 Dec 03 13:26:35 crc kubenswrapper[4690]: I1203 13:26:35.947497 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-95zhf" event={"ID":"ff4c17ee-a756-4743-aeb8-aae8fe520712","Type":"ContainerStarted","Data":"8c1dff448f18cdb06ebc376b9ed1b0108c5c645f0c5fc8ef1ff482f598cb6f71"} Dec 03 13:26:35 crc kubenswrapper[4690]: I1203 13:26:35.947535 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-95zhf" event={"ID":"ff4c17ee-a756-4743-aeb8-aae8fe520712","Type":"ContainerStarted","Data":"6c9c9e52d254001e15d2457da474aa653fa989c9f313b7a7aad4cf83bd7906e9"} Dec 03 13:26:35 crc kubenswrapper[4690]: I1203 13:26:35.974189 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pl6hw" podStartSLOduration=2.860449977 podStartE2EDuration="5.97416222s" podCreationTimestamp="2025-12-03 13:26:30 +0000 UTC" firstStartedPulling="2025-12-03 13:26:31.733577722 +0000 UTC m=+1037.714498155" lastFinishedPulling="2025-12-03 13:26:34.847289965 +0000 UTC m=+1040.828210398" observedRunningTime="2025-12-03 13:26:35.964636207 +0000 UTC m=+1041.945556650" watchObservedRunningTime="2025-12-03 13:26:35.97416222 +0000 UTC m=+1041.955082693" Dec 03 13:26:35 crc kubenswrapper[4690]: I1203 13:26:35.981305 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-95zhf" podStartSLOduration=1.894173421 podStartE2EDuration="1.981279941s" podCreationTimestamp="2025-12-03 13:26:34 +0000 UTC" firstStartedPulling="2025-12-03 13:26:35.22617521 +0000 UTC m=+1041.207095653" lastFinishedPulling="2025-12-03 13:26:35.31328174 +0000 UTC m=+1041.294202173" observedRunningTime="2025-12-03 13:26:35.979080545 +0000 UTC m=+1041.960000988" watchObservedRunningTime="2025-12-03 13:26:35.981279941 +0000 UTC m=+1041.962200374" Dec 03 13:26:36 crc kubenswrapper[4690]: I1203 13:26:36.172041 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-r8zcv" Dec 03 13:26:36 crc kubenswrapper[4690]: I1203 13:26:36.955892 4690 generic.go:334] "Generic (PLEG): container finished" podID="a58fc074-d792-4cec-87c7-b13197c39733" containerID="49dc2db42f852366b40cc26f12ebb86b15d30c3bcc8fc0624bd9b38b48914fbd" exitCode=0 Dec 03 13:26:36 crc kubenswrapper[4690]: I1203 13:26:36.956696 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pl6hw" event={"ID":"a58fc074-d792-4cec-87c7-b13197c39733","Type":"ContainerDied","Data":"49dc2db42f852366b40cc26f12ebb86b15d30c3bcc8fc0624bd9b38b48914fbd"} Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.069410 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hfmdl" Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.463082 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pl6hw" Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.563572 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v7d6\" (UniqueName: \"kubernetes.io/projected/a58fc074-d792-4cec-87c7-b13197c39733-kube-api-access-4v7d6\") pod \"a58fc074-d792-4cec-87c7-b13197c39733\" (UID: \"a58fc074-d792-4cec-87c7-b13197c39733\") " Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.578327 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a58fc074-d792-4cec-87c7-b13197c39733-kube-api-access-4v7d6" (OuterVolumeSpecName: "kube-api-access-4v7d6") pod "a58fc074-d792-4cec-87c7-b13197c39733" (UID: "a58fc074-d792-4cec-87c7-b13197c39733"). InnerVolumeSpecName "kube-api-access-4v7d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.665467 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v7d6\" (UniqueName: \"kubernetes.io/projected/a58fc074-d792-4cec-87c7-b13197c39733-kube-api-access-4v7d6\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.965296 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pl6hw" event={"ID":"a58fc074-d792-4cec-87c7-b13197c39733","Type":"ContainerDied","Data":"2906729a55d65584c0b2c34c0a528eff60b8d3b632fe56f70ba997306abca1b4"} Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.965352 4690 scope.go:117] "RemoveContainer" containerID="49dc2db42f852366b40cc26f12ebb86b15d30c3bcc8fc0624bd9b38b48914fbd" Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.965351 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pl6hw" Dec 03 13:26:37 crc kubenswrapper[4690]: I1203 13:26:37.995629 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-pl6hw"] Dec 03 13:26:38 crc kubenswrapper[4690]: I1203 13:26:38.001517 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-pl6hw"] Dec 03 13:26:38 crc kubenswrapper[4690]: I1203 13:26:38.323379 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a58fc074-d792-4cec-87c7-b13197c39733" path="/var/lib/kubelet/pods/a58fc074-d792-4cec-87c7-b13197c39733/volumes" Dec 03 13:26:44 crc kubenswrapper[4690]: I1203 13:26:44.781061 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:44 crc kubenswrapper[4690]: I1203 13:26:44.781683 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:44 crc kubenswrapper[4690]: I1203 13:26:44.807051 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:45 crc kubenswrapper[4690]: I1203 13:26:45.036806 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-95zhf" Dec 03 13:26:46 crc kubenswrapper[4690]: I1203 13:26:46.823668 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:26:46 crc kubenswrapper[4690]: I1203 13:26:46.824086 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:26:46 crc kubenswrapper[4690]: I1203 13:26:46.824134 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:26:46 crc kubenswrapper[4690]: I1203 13:26:46.824844 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7081acf7649c775529e892a032f5a109816cdd926c9cbd0790c89b5ae1594b9f"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:26:46 crc kubenswrapper[4690]: I1203 13:26:46.824928 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://7081acf7649c775529e892a032f5a109816cdd926c9cbd0790c89b5ae1594b9f" gracePeriod=600 Dec 03 13:26:47 crc kubenswrapper[4690]: I1203 13:26:47.027856 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="7081acf7649c775529e892a032f5a109816cdd926c9cbd0790c89b5ae1594b9f" exitCode=0 Dec 03 13:26:47 crc kubenswrapper[4690]: I1203 13:26:47.027917 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"7081acf7649c775529e892a032f5a109816cdd926c9cbd0790c89b5ae1594b9f"} Dec 03 13:26:47 crc kubenswrapper[4690]: I1203 13:26:47.027953 4690 scope.go:117] "RemoveContainer" containerID="6757d3773013b08b3fe1dc029582fe576a1dfe2c310afb7b61e2a8c64d0aa40f" Dec 03 13:26:48 crc kubenswrapper[4690]: I1203 13:26:48.039797 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"22570d41f66f8cf71e73ec062157689b8763204a45407561d1852ebc8cede0d1"} Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.029349 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz"] Dec 03 13:26:52 crc kubenswrapper[4690]: E1203 13:26:52.030400 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a58fc074-d792-4cec-87c7-b13197c39733" containerName="registry-server" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.030417 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a58fc074-d792-4cec-87c7-b13197c39733" containerName="registry-server" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.030643 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a58fc074-d792-4cec-87c7-b13197c39733" containerName="registry-server" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.032045 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.035171 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6nzgt" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.042931 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz"] Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.189634 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.189749 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.189821 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnpg4\" (UniqueName: \"kubernetes.io/projected/27e2f6d9-803d-48e7-ac24-bc3093175534-kube-api-access-gnpg4\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.291044 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnpg4\" (UniqueName: \"kubernetes.io/projected/27e2f6d9-803d-48e7-ac24-bc3093175534-kube-api-access-gnpg4\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.291174 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.291252 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.291774 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.291937 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.312572 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnpg4\" (UniqueName: \"kubernetes.io/projected/27e2f6d9-803d-48e7-ac24-bc3093175534-kube-api-access-gnpg4\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.363577 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:52 crc kubenswrapper[4690]: I1203 13:26:52.793921 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz"] Dec 03 13:26:52 crc kubenswrapper[4690]: W1203 13:26:52.803301 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27e2f6d9_803d_48e7_ac24_bc3093175534.slice/crio-f5654967bb96aae88c10123c9057ce618a3535564621ebd41e0bcfd217abd229 WatchSource:0}: Error finding container f5654967bb96aae88c10123c9057ce618a3535564621ebd41e0bcfd217abd229: Status 404 returned error can't find the container with id f5654967bb96aae88c10123c9057ce618a3535564621ebd41e0bcfd217abd229 Dec 03 13:26:53 crc kubenswrapper[4690]: I1203 13:26:53.076306 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" event={"ID":"27e2f6d9-803d-48e7-ac24-bc3093175534","Type":"ContainerStarted","Data":"f48c2d3d6eefc4cde804753dc1edd914e7995c1fbeafb9f64e3dd306f8c1bb9d"} Dec 03 13:26:53 crc kubenswrapper[4690]: I1203 13:26:53.076722 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" event={"ID":"27e2f6d9-803d-48e7-ac24-bc3093175534","Type":"ContainerStarted","Data":"f5654967bb96aae88c10123c9057ce618a3535564621ebd41e0bcfd217abd229"} Dec 03 13:26:54 crc kubenswrapper[4690]: I1203 13:26:54.088949 4690 generic.go:334] "Generic (PLEG): container finished" podID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerID="f48c2d3d6eefc4cde804753dc1edd914e7995c1fbeafb9f64e3dd306f8c1bb9d" exitCode=0 Dec 03 13:26:54 crc kubenswrapper[4690]: I1203 13:26:54.089021 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" event={"ID":"27e2f6d9-803d-48e7-ac24-bc3093175534","Type":"ContainerDied","Data":"f48c2d3d6eefc4cde804753dc1edd914e7995c1fbeafb9f64e3dd306f8c1bb9d"} Dec 03 13:26:55 crc kubenswrapper[4690]: I1203 13:26:55.099009 4690 generic.go:334] "Generic (PLEG): container finished" podID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerID="07931ca9751ca741994b05ca0c81d9b851d238b6c62f1b05f7b54197ab7a99f3" exitCode=0 Dec 03 13:26:55 crc kubenswrapper[4690]: I1203 13:26:55.099382 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" event={"ID":"27e2f6d9-803d-48e7-ac24-bc3093175534","Type":"ContainerDied","Data":"07931ca9751ca741994b05ca0c81d9b851d238b6c62f1b05f7b54197ab7a99f3"} Dec 03 13:26:56 crc kubenswrapper[4690]: I1203 13:26:56.109554 4690 generic.go:334] "Generic (PLEG): container finished" podID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerID="897db1544d2002063223de8d761193f4e2f68db141a3a96f995de7e01c2a575a" exitCode=0 Dec 03 13:26:56 crc kubenswrapper[4690]: I1203 13:26:56.109644 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" event={"ID":"27e2f6d9-803d-48e7-ac24-bc3093175534","Type":"ContainerDied","Data":"897db1544d2002063223de8d761193f4e2f68db141a3a96f995de7e01c2a575a"} Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.402250 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.475574 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnpg4\" (UniqueName: \"kubernetes.io/projected/27e2f6d9-803d-48e7-ac24-bc3093175534-kube-api-access-gnpg4\") pod \"27e2f6d9-803d-48e7-ac24-bc3093175534\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.475769 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-util\") pod \"27e2f6d9-803d-48e7-ac24-bc3093175534\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.475828 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-bundle\") pod \"27e2f6d9-803d-48e7-ac24-bc3093175534\" (UID: \"27e2f6d9-803d-48e7-ac24-bc3093175534\") " Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.476593 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-bundle" (OuterVolumeSpecName: "bundle") pod "27e2f6d9-803d-48e7-ac24-bc3093175534" (UID: "27e2f6d9-803d-48e7-ac24-bc3093175534"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.482951 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e2f6d9-803d-48e7-ac24-bc3093175534-kube-api-access-gnpg4" (OuterVolumeSpecName: "kube-api-access-gnpg4") pod "27e2f6d9-803d-48e7-ac24-bc3093175534" (UID: "27e2f6d9-803d-48e7-ac24-bc3093175534"). InnerVolumeSpecName "kube-api-access-gnpg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.490238 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-util" (OuterVolumeSpecName: "util") pod "27e2f6d9-803d-48e7-ac24-bc3093175534" (UID: "27e2f6d9-803d-48e7-ac24-bc3093175534"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.578323 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnpg4\" (UniqueName: \"kubernetes.io/projected/27e2f6d9-803d-48e7-ac24-bc3093175534-kube-api-access-gnpg4\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.578376 4690 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-util\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:57 crc kubenswrapper[4690]: I1203 13:26:57.578391 4690 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27e2f6d9-803d-48e7-ac24-bc3093175534-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:58 crc kubenswrapper[4690]: I1203 13:26:58.127658 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" event={"ID":"27e2f6d9-803d-48e7-ac24-bc3093175534","Type":"ContainerDied","Data":"f5654967bb96aae88c10123c9057ce618a3535564621ebd41e0bcfd217abd229"} Dec 03 13:26:58 crc kubenswrapper[4690]: I1203 13:26:58.127713 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5654967bb96aae88c10123c9057ce618a3535564621ebd41e0bcfd217abd229" Dec 03 13:26:58 crc kubenswrapper[4690]: I1203 13:26:58.127839 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.039816 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj"] Dec 03 13:27:04 crc kubenswrapper[4690]: E1203 13:27:04.040776 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerName="util" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.040791 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerName="util" Dec 03 13:27:04 crc kubenswrapper[4690]: E1203 13:27:04.040816 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerName="extract" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.040824 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerName="extract" Dec 03 13:27:04 crc kubenswrapper[4690]: E1203 13:27:04.040839 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerName="pull" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.040847 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerName="pull" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.041001 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e2f6d9-803d-48e7-ac24-bc3093175534" containerName="extract" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.041592 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.045623 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-6sdvt" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.121933 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj"] Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.177958 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjs62\" (UniqueName: \"kubernetes.io/projected/b3182947-b66b-4777-8410-b6ffc37aa5d0-kube-api-access-xjs62\") pod \"openstack-operator-controller-operator-58bc846945-vrqwj\" (UID: \"b3182947-b66b-4777-8410-b6ffc37aa5d0\") " pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.279346 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjs62\" (UniqueName: \"kubernetes.io/projected/b3182947-b66b-4777-8410-b6ffc37aa5d0-kube-api-access-xjs62\") pod \"openstack-operator-controller-operator-58bc846945-vrqwj\" (UID: \"b3182947-b66b-4777-8410-b6ffc37aa5d0\") " pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.301055 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjs62\" (UniqueName: \"kubernetes.io/projected/b3182947-b66b-4777-8410-b6ffc37aa5d0-kube-api-access-xjs62\") pod \"openstack-operator-controller-operator-58bc846945-vrqwj\" (UID: \"b3182947-b66b-4777-8410-b6ffc37aa5d0\") " pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.363092 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" Dec 03 13:27:04 crc kubenswrapper[4690]: I1203 13:27:04.643202 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj"] Dec 03 13:27:05 crc kubenswrapper[4690]: I1203 13:27:05.183760 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" event={"ID":"b3182947-b66b-4777-8410-b6ffc37aa5d0","Type":"ContainerStarted","Data":"9798f220b6a8da6dfa760bee6bdf9d695240eeb4ee6cc78c2d4ee4823e9f3de1"} Dec 03 13:27:10 crc kubenswrapper[4690]: I1203 13:27:10.233796 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" event={"ID":"b3182947-b66b-4777-8410-b6ffc37aa5d0","Type":"ContainerStarted","Data":"edd869a5ac47efa396fa010d7b3f8a3eac7dc80bfc24f5330c22a002b0b25070"} Dec 03 13:27:10 crc kubenswrapper[4690]: I1203 13:27:10.234509 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" Dec 03 13:27:10 crc kubenswrapper[4690]: I1203 13:27:10.262461 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" podStartSLOduration=1.211647104 podStartE2EDuration="6.262444218s" podCreationTimestamp="2025-12-03 13:27:04 +0000 UTC" firstStartedPulling="2025-12-03 13:27:04.648250227 +0000 UTC m=+1070.629170660" lastFinishedPulling="2025-12-03 13:27:09.699047341 +0000 UTC m=+1075.679967774" observedRunningTime="2025-12-03 13:27:10.261047202 +0000 UTC m=+1076.241967645" watchObservedRunningTime="2025-12-03 13:27:10.262444218 +0000 UTC m=+1076.243364651" Dec 03 13:27:14 crc kubenswrapper[4690]: I1203 13:27:14.367173 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-58bc846945-vrqwj" Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.924782 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v"] Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.926812 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.928831 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-g668f" Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.935218 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v"] Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.946352 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq"] Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.947780 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.952621 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-nnhmh" Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.961902 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq"] Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.984121 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb"] Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.985633 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" Dec 03 13:27:51 crc kubenswrapper[4690]: I1203 13:27:51.989280 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-lcrbd" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.000544 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.001808 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.006554 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-tj2hk" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.035948 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.049546 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.060192 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-662gz\" (UniqueName: \"kubernetes.io/projected/02c9e233-e41c-4fd2-9165-0ad5413fdaa2-kube-api-access-662gz\") pod \"designate-operator-controller-manager-78b4bc895b-hwxjb\" (UID: \"02c9e233-e41c-4fd2-9165-0ad5413fdaa2\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.060272 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpgsj\" (UniqueName: \"kubernetes.io/projected/1239bbbc-6613-4e1e-867b-fd62dee3c32b-kube-api-access-tpgsj\") pod \"barbican-operator-controller-manager-7d9dfd778-hx22v\" (UID: \"1239bbbc-6613-4e1e-867b-fd62dee3c32b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.060307 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dct62\" (UniqueName: \"kubernetes.io/projected/3b3afb5d-aef4-4c05-876f-52fa03532fe1-kube-api-access-dct62\") pod \"glance-operator-controller-manager-77987cd8cd-jmz4q\" (UID: \"3b3afb5d-aef4-4c05-876f-52fa03532fe1\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.060361 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9q5b\" (UniqueName: \"kubernetes.io/projected/bc776d95-7342-47e3-b2ad-abc4a3f7aa23-kube-api-access-v9q5b\") pod \"cinder-operator-controller-manager-859b6ccc6-pnhhq\" (UID: \"bc776d95-7342-47e3-b2ad-abc4a3f7aa23\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.060452 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.061702 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.067069 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-6fkdg" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.072822 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.075376 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.077604 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-dl9lw" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.094282 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.116115 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.139228 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.140437 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.144393 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.146726 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.148137 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.148614 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4dh6l" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.154407 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-jk46c" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.162897 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-662gz\" (UniqueName: \"kubernetes.io/projected/02c9e233-e41c-4fd2-9165-0ad5413fdaa2-kube-api-access-662gz\") pod \"designate-operator-controller-manager-78b4bc895b-hwxjb\" (UID: \"02c9e233-e41c-4fd2-9165-0ad5413fdaa2\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.163150 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpgsj\" (UniqueName: \"kubernetes.io/projected/1239bbbc-6613-4e1e-867b-fd62dee3c32b-kube-api-access-tpgsj\") pod \"barbican-operator-controller-manager-7d9dfd778-hx22v\" (UID: \"1239bbbc-6613-4e1e-867b-fd62dee3c32b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.163274 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dct62\" (UniqueName: \"kubernetes.io/projected/3b3afb5d-aef4-4c05-876f-52fa03532fe1-kube-api-access-dct62\") pod \"glance-operator-controller-manager-77987cd8cd-jmz4q\" (UID: \"3b3afb5d-aef4-4c05-876f-52fa03532fe1\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.163383 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbxsx\" (UniqueName: \"kubernetes.io/projected/4a193df2-0e41-48f0-a997-6f4b4e65bfe6-kube-api-access-gbxsx\") pod \"horizon-operator-controller-manager-68c6d99b8f-gtcwg\" (UID: \"4a193df2-0e41-48f0-a997-6f4b4e65bfe6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.163497 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxwnw\" (UniqueName: \"kubernetes.io/projected/9da3821b-8edf-46c2-ac1c-5cabd4124968-kube-api-access-qxwnw\") pod \"heat-operator-controller-manager-5f64f6f8bb-k6ht5\" (UID: \"9da3821b-8edf-46c2-ac1c-5cabd4124968\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.163665 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9q5b\" (UniqueName: \"kubernetes.io/projected/bc776d95-7342-47e3-b2ad-abc4a3f7aa23-kube-api-access-v9q5b\") pod \"cinder-operator-controller-manager-859b6ccc6-pnhhq\" (UID: \"bc776d95-7342-47e3-b2ad-abc4a3f7aa23\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.224981 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.231255 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpgsj\" (UniqueName: \"kubernetes.io/projected/1239bbbc-6613-4e1e-867b-fd62dee3c32b-kube-api-access-tpgsj\") pod \"barbican-operator-controller-manager-7d9dfd778-hx22v\" (UID: \"1239bbbc-6613-4e1e-867b-fd62dee3c32b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.233658 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dct62\" (UniqueName: \"kubernetes.io/projected/3b3afb5d-aef4-4c05-876f-52fa03532fe1-kube-api-access-dct62\") pod \"glance-operator-controller-manager-77987cd8cd-jmz4q\" (UID: \"3b3afb5d-aef4-4c05-876f-52fa03532fe1\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.234166 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9q5b\" (UniqueName: \"kubernetes.io/projected/bc776d95-7342-47e3-b2ad-abc4a3f7aa23-kube-api-access-v9q5b\") pod \"cinder-operator-controller-manager-859b6ccc6-pnhhq\" (UID: \"bc776d95-7342-47e3-b2ad-abc4a3f7aa23\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.247624 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.264450 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-662gz\" (UniqueName: \"kubernetes.io/projected/02c9e233-e41c-4fd2-9165-0ad5413fdaa2-kube-api-access-662gz\") pod \"designate-operator-controller-manager-78b4bc895b-hwxjb\" (UID: \"02c9e233-e41c-4fd2-9165-0ad5413fdaa2\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.275716 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbxsx\" (UniqueName: \"kubernetes.io/projected/4a193df2-0e41-48f0-a997-6f4b4e65bfe6-kube-api-access-gbxsx\") pod \"horizon-operator-controller-manager-68c6d99b8f-gtcwg\" (UID: \"4a193df2-0e41-48f0-a997-6f4b4e65bfe6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.275794 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxwnw\" (UniqueName: \"kubernetes.io/projected/9da3821b-8edf-46c2-ac1c-5cabd4124968-kube-api-access-qxwnw\") pod \"heat-operator-controller-manager-5f64f6f8bb-k6ht5\" (UID: \"9da3821b-8edf-46c2-ac1c-5cabd4124968\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.275849 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf6cw\" (UniqueName: \"kubernetes.io/projected/5f4f6cc8-0d38-4c2a-b4f5-10a870c48479-kube-api-access-tf6cw\") pod \"ironic-operator-controller-manager-6c548fd776-kshv8\" (UID: \"5f4f6cc8-0d38-4c2a-b4f5-10a870c48479\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.276107 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.276219 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-785sv\" (UniqueName: \"kubernetes.io/projected/729e315d-2bf3-4568-941e-98ef7ad281eb-kube-api-access-785sv\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.280981 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.282551 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.320372 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbxsx\" (UniqueName: \"kubernetes.io/projected/4a193df2-0e41-48f0-a997-6f4b4e65bfe6-kube-api-access-gbxsx\") pod \"horizon-operator-controller-manager-68c6d99b8f-gtcwg\" (UID: \"4a193df2-0e41-48f0-a997-6f4b4e65bfe6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.333487 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.342849 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxwnw\" (UniqueName: \"kubernetes.io/projected/9da3821b-8edf-46c2-ac1c-5cabd4124968-kube-api-access-qxwnw\") pod \"heat-operator-controller-manager-5f64f6f8bb-k6ht5\" (UID: \"9da3821b-8edf-46c2-ac1c-5cabd4124968\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.347543 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.351415 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.352488 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.358850 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2gf5j" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.372124 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.373482 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.378830 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-nqsh9" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.379732 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.379801 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-785sv\" (UniqueName: \"kubernetes.io/projected/729e315d-2bf3-4568-941e-98ef7ad281eb-kube-api-access-785sv\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.379896 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf6cw\" (UniqueName: \"kubernetes.io/projected/5f4f6cc8-0d38-4c2a-b4f5-10a870c48479-kube-api-access-tf6cw\") pod \"ironic-operator-controller-manager-6c548fd776-kshv8\" (UID: \"5f4f6cc8-0d38-4c2a-b4f5-10a870c48479\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" Dec 03 13:27:52 crc kubenswrapper[4690]: E1203 13:27:52.380218 4690 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 13:27:52 crc kubenswrapper[4690]: E1203 13:27:52.380270 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert podName:729e315d-2bf3-4568-941e-98ef7ad281eb nodeName:}" failed. No retries permitted until 2025-12-03 13:27:52.880253586 +0000 UTC m=+1118.861174019 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert") pod "infra-operator-controller-manager-57548d458d-rrqhz" (UID: "729e315d-2bf3-4568-941e-98ef7ad281eb") : secret "infra-operator-webhook-server-cert" not found Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.381823 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.383366 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.385306 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7rccz" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.386365 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.394137 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.398990 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.400489 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-785sv\" (UniqueName: \"kubernetes.io/projected/729e315d-2bf3-4568-941e-98ef7ad281eb-kube-api-access-785sv\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.410907 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.419718 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.421097 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.421150 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.424313 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf6cw\" (UniqueName: \"kubernetes.io/projected/5f4f6cc8-0d38-4c2a-b4f5-10a870c48479-kube-api-access-tf6cw\") pod \"ironic-operator-controller-manager-6c548fd776-kshv8\" (UID: \"5f4f6cc8-0d38-4c2a-b4f5-10a870c48479\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.436574 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.441111 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.450423 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-6mp4f" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.451111 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-7xk2q" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.488053 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.489678 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.491091 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g24d6\" (UniqueName: \"kubernetes.io/projected/942df52e-3f08-47ef-bbeb-fe1191ced3e0-kube-api-access-g24d6\") pod \"nova-operator-controller-manager-697bc559fc-5fll5\" (UID: \"942df52e-3f08-47ef-bbeb-fe1191ced3e0\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.491135 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k82zt\" (UniqueName: \"kubernetes.io/projected/c731a6c0-1c65-45e1-893d-f620e83b5363-kube-api-access-k82zt\") pod \"manila-operator-controller-manager-7c79b5df47-jk4h5\" (UID: \"c731a6c0-1c65-45e1-893d-f620e83b5363\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.491221 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bf47\" (UniqueName: \"kubernetes.io/projected/0e42df6f-0109-4a0b-afeb-630bff5b8c35-kube-api-access-8bf47\") pod \"keystone-operator-controller-manager-7765d96ddf-49f9d\" (UID: \"0e42df6f-0109-4a0b-afeb-630bff5b8c35\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.491246 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bssm\" (UniqueName: \"kubernetes.io/projected/abb876f0-132b-4224-a3a7-074beba3367b-kube-api-access-7bssm\") pod \"mariadb-operator-controller-manager-56bbcc9d85-6g8xn\" (UID: \"abb876f0-132b-4224-a3a7-074beba3367b\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.493292 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-nmfl2" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.498929 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.491281 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgf9g\" (UniqueName: \"kubernetes.io/projected/87cb4391-ed76-4f7a-a4b5-481d409bba7d-kube-api-access-dgf9g\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-p559w\" (UID: \"87cb4391-ed76-4f7a-a4b5-481d409bba7d\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.520713 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.523560 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.533738 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.535459 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.540809 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-vlxsl" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.540998 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.573810 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.618952 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.620320 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.622716 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-f957v" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.631017 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.635047 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g24d6\" (UniqueName: \"kubernetes.io/projected/942df52e-3f08-47ef-bbeb-fe1191ced3e0-kube-api-access-g24d6\") pod \"nova-operator-controller-manager-697bc559fc-5fll5\" (UID: \"942df52e-3f08-47ef-bbeb-fe1191ced3e0\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.635111 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k82zt\" (UniqueName: \"kubernetes.io/projected/c731a6c0-1c65-45e1-893d-f620e83b5363-kube-api-access-k82zt\") pod \"manila-operator-controller-manager-7c79b5df47-jk4h5\" (UID: \"c731a6c0-1c65-45e1-893d-f620e83b5363\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.643027 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f62w5\" (UniqueName: \"kubernetes.io/projected/9890f67c-68ed-475b-9089-262d1fc270b4-kube-api-access-f62w5\") pod \"octavia-operator-controller-manager-998648c74-6q5ss\" (UID: \"9890f67c-68ed-475b-9089-262d1fc270b4\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.643184 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmvh7\" (UniqueName: \"kubernetes.io/projected/f8b73cbf-3a07-46fa-98a4-db492ee1b711-kube-api-access-gmvh7\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.643208 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.643285 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bf47\" (UniqueName: \"kubernetes.io/projected/0e42df6f-0109-4a0b-afeb-630bff5b8c35-kube-api-access-8bf47\") pod \"keystone-operator-controller-manager-7765d96ddf-49f9d\" (UID: \"0e42df6f-0109-4a0b-afeb-630bff5b8c35\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.643306 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bssm\" (UniqueName: \"kubernetes.io/projected/abb876f0-132b-4224-a3a7-074beba3367b-kube-api-access-7bssm\") pod \"mariadb-operator-controller-manager-56bbcc9d85-6g8xn\" (UID: \"abb876f0-132b-4224-a3a7-074beba3367b\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.643358 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgf9g\" (UniqueName: \"kubernetes.io/projected/87cb4391-ed76-4f7a-a4b5-481d409bba7d-kube-api-access-dgf9g\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-p559w\" (UID: \"87cb4391-ed76-4f7a-a4b5-481d409bba7d\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.676602 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k82zt\" (UniqueName: \"kubernetes.io/projected/c731a6c0-1c65-45e1-893d-f620e83b5363-kube-api-access-k82zt\") pod \"manila-operator-controller-manager-7c79b5df47-jk4h5\" (UID: \"c731a6c0-1c65-45e1-893d-f620e83b5363\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.676680 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.678307 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.687816 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g24d6\" (UniqueName: \"kubernetes.io/projected/942df52e-3f08-47ef-bbeb-fe1191ced3e0-kube-api-access-g24d6\") pod \"nova-operator-controller-manager-697bc559fc-5fll5\" (UID: \"942df52e-3f08-47ef-bbeb-fe1191ced3e0\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.702994 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-ll5qp" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.708122 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.721978 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.728092 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.736766 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.736933 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.744332 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bf47\" (UniqueName: \"kubernetes.io/projected/0e42df6f-0109-4a0b-afeb-630bff5b8c35-kube-api-access-8bf47\") pod \"keystone-operator-controller-manager-7765d96ddf-49f9d\" (UID: \"0e42df6f-0109-4a0b-afeb-630bff5b8c35\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.744572 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-t2pb8" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.745591 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x5lk\" (UniqueName: \"kubernetes.io/projected/7ef3f1c5-17d7-42f0-949b-1dbab5568051-kube-api-access-4x5lk\") pod \"ovn-operator-controller-manager-b6456fdb6-lmq2n\" (UID: \"7ef3f1c5-17d7-42f0-949b-1dbab5568051\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.745644 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf6c2\" (UniqueName: \"kubernetes.io/projected/55e64346-52e8-49b2-81de-057a98be5150-kube-api-access-rf6c2\") pod \"placement-operator-controller-manager-78f8948974-7xsmn\" (UID: \"55e64346-52e8-49b2-81de-057a98be5150\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.745692 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f62w5\" (UniqueName: \"kubernetes.io/projected/9890f67c-68ed-475b-9089-262d1fc270b4-kube-api-access-f62w5\") pod \"octavia-operator-controller-manager-998648c74-6q5ss\" (UID: \"9890f67c-68ed-475b-9089-262d1fc270b4\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.745755 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmvh7\" (UniqueName: \"kubernetes.io/projected/f8b73cbf-3a07-46fa-98a4-db492ee1b711-kube-api-access-gmvh7\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.745779 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:52 crc kubenswrapper[4690]: E1203 13:27:52.745943 4690 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.745954 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgf9g\" (UniqueName: \"kubernetes.io/projected/87cb4391-ed76-4f7a-a4b5-481d409bba7d-kube-api-access-dgf9g\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-p559w\" (UID: \"87cb4391-ed76-4f7a-a4b5-481d409bba7d\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" Dec 03 13:27:52 crc kubenswrapper[4690]: E1203 13:27:52.745986 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert podName:f8b73cbf-3a07-46fa-98a4-db492ee1b711 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:53.245971908 +0000 UTC m=+1119.226892341 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" (UID: "f8b73cbf-3a07-46fa-98a4-db492ee1b711") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.746806 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bssm\" (UniqueName: \"kubernetes.io/projected/abb876f0-132b-4224-a3a7-074beba3367b-kube-api-access-7bssm\") pod \"mariadb-operator-controller-manager-56bbcc9d85-6g8xn\" (UID: \"abb876f0-132b-4224-a3a7-074beba3367b\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.749512 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.750761 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.758611 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.759171 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.760644 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-tw4gb" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.764533 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.793543 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f62w5\" (UniqueName: \"kubernetes.io/projected/9890f67c-68ed-475b-9089-262d1fc270b4-kube-api-access-f62w5\") pod \"octavia-operator-controller-manager-998648c74-6q5ss\" (UID: \"9890f67c-68ed-475b-9089-262d1fc270b4\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.793882 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmvh7\" (UniqueName: \"kubernetes.io/projected/f8b73cbf-3a07-46fa-98a4-db492ee1b711-kube-api-access-gmvh7\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.811851 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.852550 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf6c2\" (UniqueName: \"kubernetes.io/projected/55e64346-52e8-49b2-81de-057a98be5150-kube-api-access-rf6c2\") pod \"placement-operator-controller-manager-78f8948974-7xsmn\" (UID: \"55e64346-52e8-49b2-81de-057a98be5150\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.852746 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5msqm\" (UniqueName: \"kubernetes.io/projected/7df25165-6b23-4dda-8105-05a933bc3ac8-kube-api-access-5msqm\") pod \"telemetry-operator-controller-manager-d8d6bf864-kzk67\" (UID: \"7df25165-6b23-4dda-8105-05a933bc3ac8\") " pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.852823 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddbjv\" (UniqueName: \"kubernetes.io/projected/00aeb5e7-0973-4852-9fcf-6a51c3ef9720-kube-api-access-ddbjv\") pod \"swift-operator-controller-manager-5f8c65bbfc-6xs6t\" (UID: \"00aeb5e7-0973-4852-9fcf-6a51c3ef9720\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.853071 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x5lk\" (UniqueName: \"kubernetes.io/projected/7ef3f1c5-17d7-42f0-949b-1dbab5568051-kube-api-access-4x5lk\") pod \"ovn-operator-controller-manager-b6456fdb6-lmq2n\" (UID: \"7ef3f1c5-17d7-42f0-949b-1dbab5568051\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.871215 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.874716 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.876412 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.880261 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-5nmns" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.881470 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.883305 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf6c2\" (UniqueName: \"kubernetes.io/projected/55e64346-52e8-49b2-81de-057a98be5150-kube-api-access-rf6c2\") pod \"placement-operator-controller-manager-78f8948974-7xsmn\" (UID: \"55e64346-52e8-49b2-81de-057a98be5150\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.886833 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x5lk\" (UniqueName: \"kubernetes.io/projected/7ef3f1c5-17d7-42f0-949b-1dbab5568051-kube-api-access-4x5lk\") pod \"ovn-operator-controller-manager-b6456fdb6-lmq2n\" (UID: \"7ef3f1c5-17d7-42f0-949b-1dbab5568051\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.959558 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv6tw\" (UniqueName: \"kubernetes.io/projected/0e2ce38d-747d-408f-8768-48a7310c0f9c-kube-api-access-pv6tw\") pod \"test-operator-controller-manager-5854674fcc-x7xbr\" (UID: \"0e2ce38d-747d-408f-8768-48a7310c0f9c\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.959682 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5msqm\" (UniqueName: \"kubernetes.io/projected/7df25165-6b23-4dda-8105-05a933bc3ac8-kube-api-access-5msqm\") pod \"telemetry-operator-controller-manager-d8d6bf864-kzk67\" (UID: \"7df25165-6b23-4dda-8105-05a933bc3ac8\") " pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.959757 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddbjv\" (UniqueName: \"kubernetes.io/projected/00aeb5e7-0973-4852-9fcf-6a51c3ef9720-kube-api-access-ddbjv\") pod \"swift-operator-controller-manager-5f8c65bbfc-6xs6t\" (UID: \"00aeb5e7-0973-4852-9fcf-6a51c3ef9720\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.960616 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:52 crc kubenswrapper[4690]: E1203 13:27:52.960833 4690 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 13:27:52 crc kubenswrapper[4690]: E1203 13:27:52.960926 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert podName:729e315d-2bf3-4568-941e-98ef7ad281eb nodeName:}" failed. No retries permitted until 2025-12-03 13:27:53.960903047 +0000 UTC m=+1119.941823480 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert") pod "infra-operator-controller-manager-57548d458d-rrqhz" (UID: "729e315d-2bf3-4568-941e-98ef7ad281eb") : secret "infra-operator-webhook-server-cert" not found Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.973009 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr"] Dec 03 13:27:52 crc kubenswrapper[4690]: I1203 13:27:52.976001 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.024252 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.026588 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddbjv\" (UniqueName: \"kubernetes.io/projected/00aeb5e7-0973-4852-9fcf-6a51c3ef9720-kube-api-access-ddbjv\") pod \"swift-operator-controller-manager-5f8c65bbfc-6xs6t\" (UID: \"00aeb5e7-0973-4852-9fcf-6a51c3ef9720\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.028297 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5msqm\" (UniqueName: \"kubernetes.io/projected/7df25165-6b23-4dda-8105-05a933bc3ac8-kube-api-access-5msqm\") pod \"telemetry-operator-controller-manager-d8d6bf864-kzk67\" (UID: \"7df25165-6b23-4dda-8105-05a933bc3ac8\") " pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.071616 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv6tw\" (UniqueName: \"kubernetes.io/projected/0e2ce38d-747d-408f-8768-48a7310c0f9c-kube-api-access-pv6tw\") pod \"test-operator-controller-manager-5854674fcc-x7xbr\" (UID: \"0e2ce38d-747d-408f-8768-48a7310c0f9c\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.113076 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.114564 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.119667 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.119699 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv6tw\" (UniqueName: \"kubernetes.io/projected/0e2ce38d-747d-408f-8768-48a7310c0f9c-kube-api-access-pv6tw\") pod \"test-operator-controller-manager-5854674fcc-x7xbr\" (UID: \"0e2ce38d-747d-408f-8768-48a7310c0f9c\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.120554 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ddq8t" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.172295 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.179471 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.184230 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thpqr\" (UniqueName: \"kubernetes.io/projected/172f574f-c4bf-450a-8da9-fb4b879275f9-kube-api-access-thpqr\") pod \"watcher-operator-controller-manager-769dc69bc-skzwp\" (UID: \"172f574f-c4bf-450a-8da9-fb4b879275f9\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.194198 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.204202 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.211073 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.213197 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.213649 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.213883 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-89kws" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.219035 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.233583 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.237390 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.239041 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.239770 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-j5hff" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.239790 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.254487 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.287903 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.287986 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thpqr\" (UniqueName: \"kubernetes.io/projected/172f574f-c4bf-450a-8da9-fb4b879275f9-kube-api-access-thpqr\") pod \"watcher-operator-controller-manager-769dc69bc-skzwp\" (UID: \"172f574f-c4bf-450a-8da9-fb4b879275f9\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.288027 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.288055 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.288111 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-757ld\" (UniqueName: \"kubernetes.io/projected/953dae05-293d-488a-96d3-af4eda42ad47-kube-api-access-757ld\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.288701 4690 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.288926 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert podName:f8b73cbf-3a07-46fa-98a4-db492ee1b711 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:54.288907493 +0000 UTC m=+1120.269827936 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" (UID: "f8b73cbf-3a07-46fa-98a4-db492ee1b711") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.313810 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thpqr\" (UniqueName: \"kubernetes.io/projected/172f574f-c4bf-450a-8da9-fb4b879275f9-kube-api-access-thpqr\") pod \"watcher-operator-controller-manager-769dc69bc-skzwp\" (UID: \"172f574f-c4bf-450a-8da9-fb4b879275f9\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.389375 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.389456 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjm42\" (UniqueName: \"kubernetes.io/projected/eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58-kube-api-access-zjm42\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fjthl\" (UID: \"eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.389503 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.389554 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-757ld\" (UniqueName: \"kubernetes.io/projected/953dae05-293d-488a-96d3-af4eda42ad47-kube-api-access-757ld\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.392295 4690 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.392364 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:53.892345776 +0000 UTC m=+1119.873266209 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "metrics-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.393595 4690 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.393689 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:53.893663849 +0000 UTC m=+1119.874584452 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "webhook-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.421098 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-757ld\" (UniqueName: \"kubernetes.io/projected/953dae05-293d-488a-96d3-af4eda42ad47-kube-api-access-757ld\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.453439 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.492168 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjm42\" (UniqueName: \"kubernetes.io/projected/eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58-kube-api-access-zjm42\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fjthl\" (UID: \"eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.517779 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjm42\" (UniqueName: \"kubernetes.io/projected/eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58-kube-api-access-zjm42\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fjthl\" (UID: \"eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.575636 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.600568 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" event={"ID":"1239bbbc-6613-4e1e-867b-fd62dee3c32b","Type":"ContainerStarted","Data":"401791414a1af648ac20b992e7a6ec80dc2cdbd7b54df0a145c0721880824e79"} Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.656620 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.665951 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq"] Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.907014 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: I1203 13:27:53.907493 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.907762 4690 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.907834 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:54.907813755 +0000 UTC m=+1120.888734188 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "webhook-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.907921 4690 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 13:27:53 crc kubenswrapper[4690]: E1203 13:27:53.907952 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:54.907940318 +0000 UTC m=+1120.888860751 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "metrics-server-cert" not found Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.008982 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.009229 4690 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.009332 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert podName:729e315d-2bf3-4568-941e-98ef7ad281eb nodeName:}" failed. No retries permitted until 2025-12-03 13:27:56.009306968 +0000 UTC m=+1121.990227551 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert") pod "infra-operator-controller-manager-57548d458d-rrqhz" (UID: "729e315d-2bf3-4568-941e-98ef7ad281eb") : secret "infra-operator-webhook-server-cert" not found Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.254956 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.266804 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.278914 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.291286 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.304175 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.328123 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.328532 4690 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.328592 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert podName:f8b73cbf-3a07-46fa-98a4-db492ee1b711 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:56.328575202 +0000 UTC m=+1122.309495635 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" (UID: "f8b73cbf-3a07-46fa-98a4-db492ee1b711") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:27:54 crc kubenswrapper[4690]: W1203 13:27:54.418194 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f4f6cc8_0d38_4c2a_b4f5_10a870c48479.slice/crio-b889b91c045481e8ed93d6a75c05bc2edd6dc514bfd696bee5098f487642c85c WatchSource:0}: Error finding container b889b91c045481e8ed93d6a75c05bc2edd6dc514bfd696bee5098f487642c85c: Status 404 returned error can't find the container with id b889b91c045481e8ed93d6a75c05bc2edd6dc514bfd696bee5098f487642c85c Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.426722 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.437329 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.472585 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5"] Dec 03 13:27:54 crc kubenswrapper[4690]: W1203 13:27:54.484625 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b3afb5d_aef4_4c05_876f_52fa03532fe1.slice/crio-8cf24914f1a7704cf195846fb64ef8771022197c81b71662e095c9a963de2e94 WatchSource:0}: Error finding container 8cf24914f1a7704cf195846fb64ef8771022197c81b71662e095c9a963de2e94: Status 404 returned error can't find the container with id 8cf24914f1a7704cf195846fb64ef8771022197c81b71662e095c9a963de2e94 Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.544572 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dct62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-jmz4q_openstack-operators(3b3afb5d-aef4-4c05-876f-52fa03532fe1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.557854 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dct62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-jmz4q_openstack-operators(3b3afb5d-aef4-4c05-876f-52fa03532fe1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.559201 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" podUID="3b3afb5d-aef4-4c05-876f-52fa03532fe1" Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.572644 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.585531 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.599097 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.638009 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" event={"ID":"02c9e233-e41c-4fd2-9165-0ad5413fdaa2","Type":"ContainerStarted","Data":"d32603c5c9fa26d58730620be0636314d08426561b64133cb5776671aeac991c"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.639247 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d"] Dec 03 13:27:54 crc kubenswrapper[4690]: W1203 13:27:54.646465 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00aeb5e7_0973_4852_9fcf_6a51c3ef9720.slice/crio-dc59453a501b3023bb2b8d3aa8db7c5db970199182ee23870a79582dc73143c6 WatchSource:0}: Error finding container dc59453a501b3023bb2b8d3aa8db7c5db970199182ee23870a79582dc73143c6: Status 404 returned error can't find the container with id dc59453a501b3023bb2b8d3aa8db7c5db970199182ee23870a79582dc73143c6 Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.652810 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" event={"ID":"87cb4391-ed76-4f7a-a4b5-481d409bba7d","Type":"ContainerStarted","Data":"62066a47ed93a578bda3a29a39cdab649d93a182c645ecbea46b11ea4dfb433f"} Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.657656 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddbjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6xs6t_openstack-operators(00aeb5e7-0973-4852-9fcf-6a51c3ef9720): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.657668 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-thpqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-skzwp_openstack-operators(172f574f-c4bf-450a-8da9-fb4b879275f9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.660057 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" event={"ID":"3b3afb5d-aef4-4c05-876f-52fa03532fe1","Type":"ContainerStarted","Data":"8cf24914f1a7704cf195846fb64ef8771022197c81b71662e095c9a963de2e94"} Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.661160 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-thpqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-skzwp_openstack-operators(172f574f-c4bf-450a-8da9-fb4b879275f9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.661251 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddbjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6xs6t_openstack-operators(00aeb5e7-0973-4852-9fcf-6a51c3ef9720): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.662599 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" podUID="00aeb5e7-0973-4852-9fcf-6a51c3ef9720" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.662724 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" podUID="172f574f-c4bf-450a-8da9-fb4b879275f9" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.663523 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" podUID="3b3afb5d-aef4-4c05-876f-52fa03532fe1" Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.667948 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" event={"ID":"abb876f0-132b-4224-a3a7-074beba3367b","Type":"ContainerStarted","Data":"cc5b8baec408fbcfa726a2677d720f1e3c5df1a8d67e77c87fcacde93f109245"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.674187 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" event={"ID":"0e42df6f-0109-4a0b-afeb-630bff5b8c35","Type":"ContainerStarted","Data":"187df296186458dd1ac0791eeb73080d8e8885e50bb5bac178c8522e7ba0e149"} Dec 03 13:27:54 crc kubenswrapper[4690]: W1203 13:27:54.675311 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7df25165_6b23_4dda_8105_05a933bc3ac8.slice/crio-035ae04947374635f14b3cb24ff297f43eacf60da4187eb634695e26edfdb6b6 WatchSource:0}: Error finding container 035ae04947374635f14b3cb24ff297f43eacf60da4187eb634695e26edfdb6b6: Status 404 returned error can't find the container with id 035ae04947374635f14b3cb24ff297f43eacf60da4187eb634695e26edfdb6b6 Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.676133 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" event={"ID":"bc776d95-7342-47e3-b2ad-abc4a3f7aa23","Type":"ContainerStarted","Data":"d227ecd32c3b0eb7af3fa0aebe8ae65599af701359b0b5d078842bf330402433"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.677180 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" event={"ID":"9890f67c-68ed-475b-9089-262d1fc270b4","Type":"ContainerStarted","Data":"25784b7d63de5ebfb1c416405477ee337e8b88006620aba4b115e3a90702cbe4"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.678071 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" event={"ID":"0e2ce38d-747d-408f-8768-48a7310c0f9c","Type":"ContainerStarted","Data":"8801e8cb61e5c8b8c1ff8fb94c14f8f86079f78066ea610fb679ef696e1570f0"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.679084 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" event={"ID":"5f4f6cc8-0d38-4c2a-b4f5-10a870c48479","Type":"ContainerStarted","Data":"b889b91c045481e8ed93d6a75c05bc2edd6dc514bfd696bee5098f487642c85c"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.681369 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" event={"ID":"942df52e-3f08-47ef-bbeb-fe1191ced3e0","Type":"ContainerStarted","Data":"885ee77adbb64791e8945dff24b932951b21db34d45fff8d7f2398296f461c47"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.682212 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" event={"ID":"c731a6c0-1c65-45e1-893d-f620e83b5363","Type":"ContainerStarted","Data":"aa81b69b1874fa00be617d7fd55891c3e841eaa3692973da4c3d171884492b8e"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.683709 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" event={"ID":"7ef3f1c5-17d7-42f0-949b-1dbab5568051","Type":"ContainerStarted","Data":"c79afe336c4ff577020cb8b55e56976867134940258bfe9a835fb2c7fa8aae09"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.685374 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" event={"ID":"9da3821b-8edf-46c2-ac1c-5cabd4124968","Type":"ContainerStarted","Data":"c29214489c48fb5a5861f97ff4ecb0f9032f7d83f6473167a33ce3d8bda6079d"} Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.686792 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" event={"ID":"4a193df2-0e41-48f0-a997-6f4b4e65bfe6","Type":"ContainerStarted","Data":"30afb310d00a86118642cb03766adadb9052a08ad8d9f24ac20c61c63afee69d"} Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.694544 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.115:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5msqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d8d6bf864-kzk67_openstack-operators(7df25165-6b23-4dda-8105-05a933bc3ac8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: W1203 13:27:54.696029 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb55cfd9_ad7f_473f_bbc6_dbb7bc436a58.slice/crio-347ae4b3d65c1240c44cbefb0b0d56b429bf9a68ee686e7d9ed6f173940a1c39 WatchSource:0}: Error finding container 347ae4b3d65c1240c44cbefb0b0d56b429bf9a68ee686e7d9ed6f173940a1c39: Status 404 returned error can't find the container with id 347ae4b3d65c1240c44cbefb0b0d56b429bf9a68ee686e7d9ed6f173940a1c39 Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.696847 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5msqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d8d6bf864-kzk67_openstack-operators(7df25165-6b23-4dda-8105-05a933bc3ac8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.698114 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" podUID="7df25165-6b23-4dda-8105-05a933bc3ac8" Dec 03 13:27:54 crc kubenswrapper[4690]: W1203 13:27:54.709111 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55e64346_52e8_49b2_81de_057a98be5150.slice/crio-75be414211cce8c58d1a269d80a0da5fc77d72b12968ed771b94de952907f3b1 WatchSource:0}: Error finding container 75be414211cce8c58d1a269d80a0da5fc77d72b12968ed771b94de952907f3b1: Status 404 returned error can't find the container with id 75be414211cce8c58d1a269d80a0da5fc77d72b12968ed771b94de952907f3b1 Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.716304 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rf6c2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-7xsmn_openstack-operators(55e64346-52e8-49b2-81de-057a98be5150): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.719181 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rf6c2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-7xsmn_openstack-operators(55e64346-52e8-49b2-81de-057a98be5150): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.721558 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" podUID="55e64346-52e8-49b2-81de-057a98be5150" Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.770502 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.778704 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.793856 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.804251 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.834653 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn"] Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.952214 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:54 crc kubenswrapper[4690]: I1203 13:27:54.952302 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.952482 4690 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.952538 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:56.95252029 +0000 UTC m=+1122.933440723 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "webhook-server-cert" not found Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.952933 4690 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 13:27:54 crc kubenswrapper[4690]: E1203 13:27:54.952961 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:27:56.952953141 +0000 UTC m=+1122.933873574 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "metrics-server-cert" not found Dec 03 13:27:55 crc kubenswrapper[4690]: I1203 13:27:55.701429 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" event={"ID":"55e64346-52e8-49b2-81de-057a98be5150","Type":"ContainerStarted","Data":"75be414211cce8c58d1a269d80a0da5fc77d72b12968ed771b94de952907f3b1"} Dec 03 13:27:55 crc kubenswrapper[4690]: I1203 13:27:55.704495 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" event={"ID":"00aeb5e7-0973-4852-9fcf-6a51c3ef9720","Type":"ContainerStarted","Data":"dc59453a501b3023bb2b8d3aa8db7c5db970199182ee23870a79582dc73143c6"} Dec 03 13:27:55 crc kubenswrapper[4690]: E1203 13:27:55.705520 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" podUID="55e64346-52e8-49b2-81de-057a98be5150" Dec 03 13:27:55 crc kubenswrapper[4690]: I1203 13:27:55.706405 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" event={"ID":"eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58","Type":"ContainerStarted","Data":"347ae4b3d65c1240c44cbefb0b0d56b429bf9a68ee686e7d9ed6f173940a1c39"} Dec 03 13:27:55 crc kubenswrapper[4690]: E1203 13:27:55.706635 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" podUID="00aeb5e7-0973-4852-9fcf-6a51c3ef9720" Dec 03 13:27:55 crc kubenswrapper[4690]: I1203 13:27:55.714084 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" event={"ID":"7df25165-6b23-4dda-8105-05a933bc3ac8","Type":"ContainerStarted","Data":"035ae04947374635f14b3cb24ff297f43eacf60da4187eb634695e26edfdb6b6"} Dec 03 13:27:55 crc kubenswrapper[4690]: E1203 13:27:55.723651 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.115:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" podUID="7df25165-6b23-4dda-8105-05a933bc3ac8" Dec 03 13:27:55 crc kubenswrapper[4690]: I1203 13:27:55.733067 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" event={"ID":"172f574f-c4bf-450a-8da9-fb4b879275f9","Type":"ContainerStarted","Data":"6ed3ac3ca5e4a34ae6ef269e121088e03aade33fa8cc718a9784b48a5ea6396c"} Dec 03 13:27:55 crc kubenswrapper[4690]: E1203 13:27:55.740576 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" podUID="3b3afb5d-aef4-4c05-876f-52fa03532fe1" Dec 03 13:27:55 crc kubenswrapper[4690]: E1203 13:27:55.749487 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" podUID="172f574f-c4bf-450a-8da9-fb4b879275f9" Dec 03 13:27:56 crc kubenswrapper[4690]: I1203 13:27:56.085408 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:27:56 crc kubenswrapper[4690]: E1203 13:27:56.085581 4690 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 13:27:56 crc kubenswrapper[4690]: E1203 13:27:56.085639 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert podName:729e315d-2bf3-4568-941e-98ef7ad281eb nodeName:}" failed. No retries permitted until 2025-12-03 13:28:00.085623378 +0000 UTC m=+1126.066543811 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert") pod "infra-operator-controller-manager-57548d458d-rrqhz" (UID: "729e315d-2bf3-4568-941e-98ef7ad281eb") : secret "infra-operator-webhook-server-cert" not found Dec 03 13:27:56 crc kubenswrapper[4690]: I1203 13:27:56.390418 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:27:56 crc kubenswrapper[4690]: E1203 13:27:56.390688 4690 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:27:56 crc kubenswrapper[4690]: E1203 13:27:56.390799 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert podName:f8b73cbf-3a07-46fa-98a4-db492ee1b711 nodeName:}" failed. No retries permitted until 2025-12-03 13:28:00.390772135 +0000 UTC m=+1126.371692568 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" (UID: "f8b73cbf-3a07-46fa-98a4-db492ee1b711") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:27:56 crc kubenswrapper[4690]: E1203 13:27:56.749707 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" podUID="00aeb5e7-0973-4852-9fcf-6a51c3ef9720" Dec 03 13:27:56 crc kubenswrapper[4690]: E1203 13:27:56.751624 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" podUID="55e64346-52e8-49b2-81de-057a98be5150" Dec 03 13:27:56 crc kubenswrapper[4690]: E1203 13:27:56.758224 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.115:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" podUID="7df25165-6b23-4dda-8105-05a933bc3ac8" Dec 03 13:27:56 crc kubenswrapper[4690]: E1203 13:27:56.758370 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" podUID="172f574f-c4bf-450a-8da9-fb4b879275f9" Dec 03 13:27:57 crc kubenswrapper[4690]: I1203 13:27:57.016995 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:57 crc kubenswrapper[4690]: I1203 13:27:57.017178 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:27:57 crc kubenswrapper[4690]: E1203 13:27:57.017335 4690 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 13:27:57 crc kubenswrapper[4690]: E1203 13:27:57.017402 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:28:01.017382431 +0000 UTC m=+1126.998302864 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "metrics-server-cert" not found Dec 03 13:27:57 crc kubenswrapper[4690]: E1203 13:27:57.017463 4690 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 13:27:57 crc kubenswrapper[4690]: E1203 13:27:57.017491 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:28:01.017482314 +0000 UTC m=+1126.998402747 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "webhook-server-cert" not found Dec 03 13:28:00 crc kubenswrapper[4690]: I1203 13:28:00.186191 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:28:00 crc kubenswrapper[4690]: E1203 13:28:00.186451 4690 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 13:28:00 crc kubenswrapper[4690]: E1203 13:28:00.186667 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert podName:729e315d-2bf3-4568-941e-98ef7ad281eb nodeName:}" failed. No retries permitted until 2025-12-03 13:28:08.186649652 +0000 UTC m=+1134.167570085 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert") pod "infra-operator-controller-manager-57548d458d-rrqhz" (UID: "729e315d-2bf3-4568-941e-98ef7ad281eb") : secret "infra-operator-webhook-server-cert" not found Dec 03 13:28:00 crc kubenswrapper[4690]: I1203 13:28:00.490465 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:28:00 crc kubenswrapper[4690]: E1203 13:28:00.490795 4690 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:28:00 crc kubenswrapper[4690]: E1203 13:28:00.490848 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert podName:f8b73cbf-3a07-46fa-98a4-db492ee1b711 nodeName:}" failed. No retries permitted until 2025-12-03 13:28:08.490832384 +0000 UTC m=+1134.471752817 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" (UID: "f8b73cbf-3a07-46fa-98a4-db492ee1b711") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 13:28:01 crc kubenswrapper[4690]: I1203 13:28:01.099395 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:01 crc kubenswrapper[4690]: I1203 13:28:01.099485 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:01 crc kubenswrapper[4690]: E1203 13:28:01.099743 4690 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 13:28:01 crc kubenswrapper[4690]: E1203 13:28:01.099802 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:28:09.099785383 +0000 UTC m=+1135.080705816 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "webhook-server-cert" not found Dec 03 13:28:01 crc kubenswrapper[4690]: E1203 13:28:01.100220 4690 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 13:28:01 crc kubenswrapper[4690]: E1203 13:28:01.100257 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs podName:953dae05-293d-488a-96d3-af4eda42ad47 nodeName:}" failed. No retries permitted until 2025-12-03 13:28:09.100246105 +0000 UTC m=+1135.081166538 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs") pod "openstack-operator-controller-manager-7874b9874-9jdm5" (UID: "953dae05-293d-488a-96d3-af4eda42ad47") : secret "metrics-server-cert" not found Dec 03 13:28:07 crc kubenswrapper[4690]: E1203 13:28:07.227259 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 03 13:28:07 crc kubenswrapper[4690]: E1203 13:28:07.228152 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k82zt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-jk4h5_openstack-operators(c731a6c0-1c65-45e1-893d-f620e83b5363): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:08 crc kubenswrapper[4690]: I1203 13:28:08.234794 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:28:08 crc kubenswrapper[4690]: I1203 13:28:08.241841 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/729e315d-2bf3-4568-941e-98ef7ad281eb-cert\") pod \"infra-operator-controller-manager-57548d458d-rrqhz\" (UID: \"729e315d-2bf3-4568-941e-98ef7ad281eb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:28:08 crc kubenswrapper[4690]: I1203 13:28:08.372548 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:28:08 crc kubenswrapper[4690]: I1203 13:28:08.541062 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:28:08 crc kubenswrapper[4690]: I1203 13:28:08.545460 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8b73cbf-3a07-46fa-98a4-db492ee1b711-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd\" (UID: \"f8b73cbf-3a07-46fa-98a4-db492ee1b711\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:28:08 crc kubenswrapper[4690]: I1203 13:28:08.808218 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:28:09 crc kubenswrapper[4690]: I1203 13:28:09.152052 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:09 crc kubenswrapper[4690]: I1203 13:28:09.152147 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:09 crc kubenswrapper[4690]: I1203 13:28:09.157123 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-metrics-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:09 crc kubenswrapper[4690]: I1203 13:28:09.161156 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/953dae05-293d-488a-96d3-af4eda42ad47-webhook-certs\") pod \"openstack-operator-controller-manager-7874b9874-9jdm5\" (UID: \"953dae05-293d-488a-96d3-af4eda42ad47\") " pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:09 crc kubenswrapper[4690]: I1203 13:28:09.373699 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:15 crc kubenswrapper[4690]: E1203 13:28:15.191922 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 03 13:28:15 crc kubenswrapper[4690]: E1203 13:28:15.192818 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7bssm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-6g8xn_openstack-operators(abb876f0-132b-4224-a3a7-074beba3367b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:15 crc kubenswrapper[4690]: E1203 13:28:15.372654 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 03 13:28:15 crc kubenswrapper[4690]: E1203 13:28:15.372877 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f62w5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-6q5ss_openstack-operators(9890f67c-68ed-475b-9089-262d1fc270b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:16 crc kubenswrapper[4690]: E1203 13:28:16.971780 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 03 13:28:16 crc kubenswrapper[4690]: E1203 13:28:16.972406 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qxwnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-k6ht5_openstack-operators(9da3821b-8edf-46c2-ac1c-5cabd4124968): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:18 crc kubenswrapper[4690]: E1203 13:28:18.009040 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 03 13:28:18 crc kubenswrapper[4690]: E1203 13:28:18.009287 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v9q5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-pnhhq_openstack-operators(bc776d95-7342-47e3-b2ad-abc4a3f7aa23): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:18 crc kubenswrapper[4690]: E1203 13:28:18.881486 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 03 13:28:18 crc kubenswrapper[4690]: E1203 13:28:18.883423 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tf6cw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-kshv8_openstack-operators(5f4f6cc8-0d38-4c2a-b4f5-10a870c48479): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:22 crc kubenswrapper[4690]: E1203 13:28:22.496797 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 13:28:22 crc kubenswrapper[4690]: E1203 13:28:22.498111 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zjm42,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-fjthl_openstack-operators(eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:22 crc kubenswrapper[4690]: E1203 13:28:22.499554 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" podUID="eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58" Dec 03 13:28:22 crc kubenswrapper[4690]: E1203 13:28:22.980483 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" podUID="eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58" Dec 03 13:28:23 crc kubenswrapper[4690]: E1203 13:28:23.113708 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 13:28:23 crc kubenswrapper[4690]: E1203 13:28:23.113997 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8bf47,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-49f9d_openstack-operators(0e42df6f-0109-4a0b-afeb-630bff5b8c35): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:23 crc kubenswrapper[4690]: E1203 13:28:23.754670 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 13:28:23 crc kubenswrapper[4690]: E1203 13:28:23.754965 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g24d6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-5fll5_openstack-operators(942df52e-3f08-47ef-bbeb-fe1191ced3e0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:29 crc kubenswrapper[4690]: E1203 13:28:29.300384 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.115:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9" Dec 03 13:28:29 crc kubenswrapper[4690]: E1203 13:28:29.301430 4690 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.115:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9" Dec 03 13:28:29 crc kubenswrapper[4690]: E1203 13:28:29.301694 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.115:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5msqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d8d6bf864-kzk67_openstack-operators(7df25165-6b23-4dda-8105-05a933bc3ac8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:28:30 crc kubenswrapper[4690]: I1203 13:28:30.127849 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5"] Dec 03 13:28:30 crc kubenswrapper[4690]: I1203 13:28:30.216104 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd"] Dec 03 13:28:30 crc kubenswrapper[4690]: I1203 13:28:30.293524 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz"] Dec 03 13:28:30 crc kubenswrapper[4690]: W1203 13:28:30.679443 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod729e315d_2bf3_4568_941e_98ef7ad281eb.slice/crio-8226af42488c4303c64f368fbbc48b6aedb1ba936c51b1cf2b8789adb1c419a3 WatchSource:0}: Error finding container 8226af42488c4303c64f368fbbc48b6aedb1ba936c51b1cf2b8789adb1c419a3: Status 404 returned error can't find the container with id 8226af42488c4303c64f368fbbc48b6aedb1ba936c51b1cf2b8789adb1c419a3 Dec 03 13:28:31 crc kubenswrapper[4690]: I1203 13:28:31.079944 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" event={"ID":"729e315d-2bf3-4568-941e-98ef7ad281eb","Type":"ContainerStarted","Data":"8226af42488c4303c64f368fbbc48b6aedb1ba936c51b1cf2b8789adb1c419a3"} Dec 03 13:28:31 crc kubenswrapper[4690]: I1203 13:28:31.090679 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" event={"ID":"87cb4391-ed76-4f7a-a4b5-481d409bba7d","Type":"ContainerStarted","Data":"5d133426297ad7b7d8104cb20260b43aa9ff86521b448800f235f8a36be83b8f"} Dec 03 13:28:31 crc kubenswrapper[4690]: I1203 13:28:31.093357 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" event={"ID":"953dae05-293d-488a-96d3-af4eda42ad47","Type":"ContainerStarted","Data":"766fccf89358a6cae07a286161cd662501834dbfa2c2475504d75687e9038789"} Dec 03 13:28:31 crc kubenswrapper[4690]: I1203 13:28:31.096671 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" event={"ID":"f8b73cbf-3a07-46fa-98a4-db492ee1b711","Type":"ContainerStarted","Data":"989e7758d3c45b2d504c209de9ff573ed40c7163d3d5e4e5a0c18c2afe1b07f3"} Dec 03 13:28:31 crc kubenswrapper[4690]: I1203 13:28:31.102389 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" event={"ID":"1239bbbc-6613-4e1e-867b-fd62dee3c32b","Type":"ContainerStarted","Data":"a0697f9b163647950ffefc4872fd069a64f880b457e1a181b40f90b8852fdd1e"} Dec 03 13:28:31 crc kubenswrapper[4690]: I1203 13:28:31.126168 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" event={"ID":"0e2ce38d-747d-408f-8768-48a7310c0f9c","Type":"ContainerStarted","Data":"4672dcd4132e988be5f4081a7ed9f294937560be9af16a96813755803123348b"} Dec 03 13:28:32 crc kubenswrapper[4690]: I1203 13:28:32.420854 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" event={"ID":"02c9e233-e41c-4fd2-9165-0ad5413fdaa2","Type":"ContainerStarted","Data":"e5c0acbc129d48846b17336400299b34cca67e8b38d9f243c6a030996d54cc01"} Dec 03 13:28:32 crc kubenswrapper[4690]: I1203 13:28:32.424279 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" event={"ID":"4a193df2-0e41-48f0-a997-6f4b4e65bfe6","Type":"ContainerStarted","Data":"d8b7973b54c6753fa216fe37479e5905a588d1ff94f00005adf3b21537731550"} Dec 03 13:28:32 crc kubenswrapper[4690]: I1203 13:28:32.428654 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" event={"ID":"7ef3f1c5-17d7-42f0-949b-1dbab5568051","Type":"ContainerStarted","Data":"892dfd81570893bbc28000f0cf92465f0e06c519fe55b0122b7d51648eddcfb0"} Dec 03 13:28:33 crc kubenswrapper[4690]: I1203 13:28:33.437810 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" event={"ID":"00aeb5e7-0973-4852-9fcf-6a51c3ef9720","Type":"ContainerStarted","Data":"c3135f32053c09e0a6f9e72f0462db468cf2fcc137f55982b7dab782e8c5a168"} Dec 03 13:28:33 crc kubenswrapper[4690]: I1203 13:28:33.439926 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" event={"ID":"172f574f-c4bf-450a-8da9-fb4b879275f9","Type":"ContainerStarted","Data":"d129a64c011954aa0d21ea199cc0b93d1575d1df5c2e80667eb1d9ccf3145382"} Dec 03 13:28:33 crc kubenswrapper[4690]: I1203 13:28:33.445018 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" event={"ID":"953dae05-293d-488a-96d3-af4eda42ad47","Type":"ContainerStarted","Data":"637cb182feeb22e9396c0e1d951939c6c4de2255fc6e8048b4cddc16bd1891e8"} Dec 03 13:28:33 crc kubenswrapper[4690]: I1203 13:28:33.445209 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:33 crc kubenswrapper[4690]: I1203 13:28:33.490511 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" podStartSLOduration=41.490483804 podStartE2EDuration="41.490483804s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:28:33.484460071 +0000 UTC m=+1159.465380514" watchObservedRunningTime="2025-12-03 13:28:33.490483804 +0000 UTC m=+1159.471404247" Dec 03 13:28:34 crc kubenswrapper[4690]: I1203 13:28:34.483079 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" event={"ID":"3b3afb5d-aef4-4c05-876f-52fa03532fe1","Type":"ContainerStarted","Data":"f9590371e1ca3b74a055b69108d240a59084bd2fcb7575bb7b85d4c72f3ea5f5"} Dec 03 13:28:34 crc kubenswrapper[4690]: I1203 13:28:34.485418 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" event={"ID":"55e64346-52e8-49b2-81de-057a98be5150","Type":"ContainerStarted","Data":"6e5a6dce9d9d67b2d7807ecd5673f65389bd8006323287e5d4c9297e053e4270"} Dec 03 13:28:39 crc kubenswrapper[4690]: I1203 13:28:39.382323 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7874b9874-9jdm5" Dec 03 13:28:44 crc kubenswrapper[4690]: E1203 13:28:44.509889 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" podUID="942df52e-3f08-47ef-bbeb-fe1191ced3e0" Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.562571 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" event={"ID":"7ef3f1c5-17d7-42f0-949b-1dbab5568051","Type":"ContainerStarted","Data":"56a57a159d944d0a3fc98f09173be33a1ead3ba67592352ece1fcd50a1afde93"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.565577 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" event={"ID":"4a193df2-0e41-48f0-a997-6f4b4e65bfe6","Type":"ContainerStarted","Data":"c1e92ee291c9f09744170917c131ffd1600a4ec2b8eae4ac2c3f9c8f3f5645c5"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.568453 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" event={"ID":"00aeb5e7-0973-4852-9fcf-6a51c3ef9720","Type":"ContainerStarted","Data":"172e69948b31315f612cf1bb62d94166daad6c8bb9d0f8e1a98579b37f8199d9"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.571404 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" event={"ID":"3b3afb5d-aef4-4c05-876f-52fa03532fe1","Type":"ContainerStarted","Data":"2d4aecd0026dafaaaaf52d9369cc8ad8ced549bca738872e6f79a7c16517103c"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.573700 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" event={"ID":"f8b73cbf-3a07-46fa-98a4-db492ee1b711","Type":"ContainerStarted","Data":"ef23dd045b48324a73fcdc8cbf495aeccb638f9af5f62539db581eacb524ca1d"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.576061 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" event={"ID":"1239bbbc-6613-4e1e-867b-fd62dee3c32b","Type":"ContainerStarted","Data":"993f8b709eefe84ee21d47ab1b12314b7546080d92aec9c163a22dfa6c5b6bcf"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.578308 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" event={"ID":"942df52e-3f08-47ef-bbeb-fe1191ced3e0","Type":"ContainerStarted","Data":"66ca1ed44580a75d79046373fdf0178de5ad95eae869dbae56622022ba3a75e0"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.582563 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" event={"ID":"0e2ce38d-747d-408f-8768-48a7310c0f9c","Type":"ContainerStarted","Data":"5f39cca53ff2f8ed91a8fda97b90913fdc1bc1a81f47bfec40fd9937c8bfdc67"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.585908 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" event={"ID":"7df25165-6b23-4dda-8105-05a933bc3ac8","Type":"ContainerStarted","Data":"a88c8791dcc876a4d63f16551a81d8bb483459631e362871cc8aadc936f25c37"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.588779 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" event={"ID":"729e315d-2bf3-4568-941e-98ef7ad281eb","Type":"ContainerStarted","Data":"d1d0a9683d1c39a8bd1e9c65c974b0b99b7b6c710a2b43021162f09cc68a6fc7"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.591745 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" event={"ID":"02c9e233-e41c-4fd2-9165-0ad5413fdaa2","Type":"ContainerStarted","Data":"c55e7966831be3ebdfc05c16d6195affe728ae759a532c3ed8bef42a71030cf3"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.594017 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" event={"ID":"9890f67c-68ed-475b-9089-262d1fc270b4","Type":"ContainerStarted","Data":"6f4c8c28948f7fbdd4c3e7f8f8439365e5ca31ca2c3c34ef6fd3a25dc4c43cb1"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.596173 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" event={"ID":"eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58","Type":"ContainerStarted","Data":"e453310feef60495103633dfc429ad5b244fb33e5f690d79804407c26fdf3eb1"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.598003 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" event={"ID":"abb876f0-132b-4224-a3a7-074beba3367b","Type":"ContainerStarted","Data":"61116d6a49924fdbe353e4edafbb4216b2e110889f469d8553de3a576690131f"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.601237 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" event={"ID":"172f574f-c4bf-450a-8da9-fb4b879275f9","Type":"ContainerStarted","Data":"a7786d7b99b6931240a005c5e3332e510fe63b02a336c6b7edae687f2383860c"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.603642 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" event={"ID":"5f4f6cc8-0d38-4c2a-b4f5-10a870c48479","Type":"ContainerStarted","Data":"7904b9dbd64a272b050457093dd19c93149b130f3dc267707ade6b1ebe8af9c9"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.605728 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" event={"ID":"bc776d95-7342-47e3-b2ad-abc4a3f7aa23","Type":"ContainerStarted","Data":"559d3f29a307da38b6db01f864c04cb1918cc88ad79159004ee97a560f527f89"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.607673 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" event={"ID":"0e42df6f-0109-4a0b-afeb-630bff5b8c35","Type":"ContainerStarted","Data":"0569bb72680fc19d0c7d549879b93184bb2235669fe29083cec98df6326421f6"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.610227 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" event={"ID":"87cb4391-ed76-4f7a-a4b5-481d409bba7d","Type":"ContainerStarted","Data":"56d2066f03aa43fc98392a2d23f1dac967e050feb76e79bb97af09f5a5bc572d"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.617407 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" event={"ID":"c731a6c0-1c65-45e1-893d-f620e83b5363","Type":"ContainerStarted","Data":"031459c79bbf9671deeb6bc013da10cc7df294285feec6d1f616074be96f6fc2"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.622715 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" event={"ID":"55e64346-52e8-49b2-81de-057a98be5150","Type":"ContainerStarted","Data":"8390443f2a361076665b9a32f64146499b60952bc7d829ef7fcd7bc64ea1a67b"} Dec 03 13:28:44 crc kubenswrapper[4690]: I1203 13:28:44.625012 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" event={"ID":"9da3821b-8edf-46c2-ac1c-5cabd4124968","Type":"ContainerStarted","Data":"34c4149288a12f8c5f2c0031e0eafb237fd63cb9de758a8fdff367a9924ea722"} Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.090501 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" podUID="7df25165-6b23-4dda-8105-05a933bc3ac8" Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.090949 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" podUID="9890f67c-68ed-475b-9089-262d1fc270b4" Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.091281 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" podUID="c731a6c0-1c65-45e1-893d-f620e83b5363" Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.094047 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" podUID="abb876f0-132b-4224-a3a7-074beba3367b" Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.094277 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" podUID="5f4f6cc8-0d38-4c2a-b4f5-10a870c48479" Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.094318 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" podUID="bc776d95-7342-47e3-b2ad-abc4a3f7aa23" Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.094611 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" podUID="0e42df6f-0109-4a0b-afeb-630bff5b8c35" Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.095612 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" podUID="9da3821b-8edf-46c2-ac1c-5cabd4124968" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.639163 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" event={"ID":"729e315d-2bf3-4568-941e-98ef7ad281eb","Type":"ContainerStarted","Data":"10fde04650742791b77e2b736dd9f3f27be3761be6e575a7a683e0bdcbfef1fb"} Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.639451 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.642755 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" event={"ID":"f8b73cbf-3a07-46fa-98a4-db492ee1b711","Type":"ContainerStarted","Data":"3bbdebe0c230dd7847d789f8d82cf8ec24f13494db7a771f01304aefb7e9aade"} Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.643760 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:28:45 crc kubenswrapper[4690]: E1203 13:28:45.653095 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.115:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" podUID="7df25165-6b23-4dda-8105-05a933bc3ac8" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.697105 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" podStartSLOduration=45.472802547 podStartE2EDuration="53.697075678s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:28:30.686805411 +0000 UTC m=+1156.667725834" lastFinishedPulling="2025-12-03 13:28:38.911078522 +0000 UTC m=+1164.891998965" observedRunningTime="2025-12-03 13:28:45.69241821 +0000 UTC m=+1171.673338653" watchObservedRunningTime="2025-12-03 13:28:45.697075678 +0000 UTC m=+1171.677996111" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.733691 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" podStartSLOduration=4.911482935 podStartE2EDuration="53.733670765s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.65748943 +0000 UTC m=+1120.638409863" lastFinishedPulling="2025-12-03 13:28:43.47967727 +0000 UTC m=+1169.460597693" observedRunningTime="2025-12-03 13:28:45.730943266 +0000 UTC m=+1171.711863709" watchObservedRunningTime="2025-12-03 13:28:45.733670765 +0000 UTC m=+1171.714591198" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.763693 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" podStartSLOduration=4.654805649 podStartE2EDuration="53.763666066s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.37148805 +0000 UTC m=+1120.352408483" lastFinishedPulling="2025-12-03 13:28:43.480348447 +0000 UTC m=+1169.461268900" observedRunningTime="2025-12-03 13:28:45.746343317 +0000 UTC m=+1171.727263750" watchObservedRunningTime="2025-12-03 13:28:45.763666066 +0000 UTC m=+1171.744586499" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.833388 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" podStartSLOduration=45.535090446 podStartE2EDuration="53.833366533s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:28:30.674418767 +0000 UTC m=+1156.655339200" lastFinishedPulling="2025-12-03 13:28:38.972694854 +0000 UTC m=+1164.953615287" observedRunningTime="2025-12-03 13:28:45.828487729 +0000 UTC m=+1171.809408162" watchObservedRunningTime="2025-12-03 13:28:45.833366533 +0000 UTC m=+1171.814286966" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.864208 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" podStartSLOduration=4.75781799 podStartE2EDuration="53.864185364s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.371905701 +0000 UTC m=+1120.352826134" lastFinishedPulling="2025-12-03 13:28:43.478273075 +0000 UTC m=+1169.459193508" observedRunningTime="2025-12-03 13:28:45.859446804 +0000 UTC m=+1171.840367257" watchObservedRunningTime="2025-12-03 13:28:45.864185364 +0000 UTC m=+1171.845105797" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.883020 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" podStartSLOduration=5.082215852 podStartE2EDuration="54.882999681s" podCreationTimestamp="2025-12-03 13:27:51 +0000 UTC" firstStartedPulling="2025-12-03 13:27:53.677756892 +0000 UTC m=+1119.658677325" lastFinishedPulling="2025-12-03 13:28:43.478540711 +0000 UTC m=+1169.459461154" observedRunningTime="2025-12-03 13:28:45.878703913 +0000 UTC m=+1171.859624346" watchObservedRunningTime="2025-12-03 13:28:45.882999681 +0000 UTC m=+1171.863920114" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.904210 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" podStartSLOduration=4.824196324 podStartE2EDuration="53.904188769s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.400099296 +0000 UTC m=+1120.381019729" lastFinishedPulling="2025-12-03 13:28:43.480091731 +0000 UTC m=+1169.461012174" observedRunningTime="2025-12-03 13:28:45.900890985 +0000 UTC m=+1171.881811428" watchObservedRunningTime="2025-12-03 13:28:45.904188769 +0000 UTC m=+1171.885109222" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.920352 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fjthl" podStartSLOduration=9.708165897 podStartE2EDuration="53.920332538s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.702094211 +0000 UTC m=+1120.683014644" lastFinishedPulling="2025-12-03 13:28:38.914260832 +0000 UTC m=+1164.895181285" observedRunningTime="2025-12-03 13:28:45.919155348 +0000 UTC m=+1171.900075781" watchObservedRunningTime="2025-12-03 13:28:45.920332538 +0000 UTC m=+1171.901252971" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.964589 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" podStartSLOduration=5.969716202 podStartE2EDuration="54.96456971s" podCreationTimestamp="2025-12-03 13:27:51 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.541176291 +0000 UTC m=+1120.522096724" lastFinishedPulling="2025-12-03 13:28:43.536029799 +0000 UTC m=+1169.516950232" observedRunningTime="2025-12-03 13:28:45.959829179 +0000 UTC m=+1171.940749612" watchObservedRunningTime="2025-12-03 13:28:45.96456971 +0000 UTC m=+1171.945490143" Dec 03 13:28:45 crc kubenswrapper[4690]: I1203 13:28:45.982516 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" podStartSLOduration=5.162172092 podStartE2EDuration="53.982502044s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.716112907 +0000 UTC m=+1120.697033340" lastFinishedPulling="2025-12-03 13:28:43.536442859 +0000 UTC m=+1169.517363292" observedRunningTime="2025-12-03 13:28:45.981575671 +0000 UTC m=+1171.962496104" watchObservedRunningTime="2025-12-03 13:28:45.982502044 +0000 UTC m=+1171.963422467" Dec 03 13:28:46 crc kubenswrapper[4690]: I1203 13:28:46.023304 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" podStartSLOduration=6.03314352 podStartE2EDuration="55.023287978s" podCreationTimestamp="2025-12-03 13:27:51 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.544383673 +0000 UTC m=+1120.525304106" lastFinishedPulling="2025-12-03 13:28:43.534528131 +0000 UTC m=+1169.515448564" observedRunningTime="2025-12-03 13:28:46.020494967 +0000 UTC m=+1172.001415410" watchObservedRunningTime="2025-12-03 13:28:46.023287978 +0000 UTC m=+1172.004208411" Dec 03 13:28:46 crc kubenswrapper[4690]: I1203 13:28:46.049119 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" podStartSLOduration=5.1744454619999996 podStartE2EDuration="54.049100683s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.65749811 +0000 UTC m=+1120.638418543" lastFinishedPulling="2025-12-03 13:28:43.532153331 +0000 UTC m=+1169.513073764" observedRunningTime="2025-12-03 13:28:46.043379438 +0000 UTC m=+1172.024299871" watchObservedRunningTime="2025-12-03 13:28:46.049100683 +0000 UTC m=+1172.030021116" Dec 03 13:28:46 crc kubenswrapper[4690]: I1203 13:28:46.082694 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" podStartSLOduration=4.879037961 podStartE2EDuration="55.082653953s" podCreationTimestamp="2025-12-03 13:27:51 +0000 UTC" firstStartedPulling="2025-12-03 13:27:53.274723134 +0000 UTC m=+1119.255643567" lastFinishedPulling="2025-12-03 13:28:43.478339126 +0000 UTC m=+1169.459259559" observedRunningTime="2025-12-03 13:28:46.0813494 +0000 UTC m=+1172.062269843" watchObservedRunningTime="2025-12-03 13:28:46.082653953 +0000 UTC m=+1172.063574386" Dec 03 13:28:46 crc kubenswrapper[4690]: I1203 13:28:46.652948 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" event={"ID":"942df52e-3f08-47ef-bbeb-fe1191ced3e0","Type":"ContainerStarted","Data":"0b6b516855149452896a55e941eb150745d006e4a30c194ac351ef886134c937"} Dec 03 13:28:46 crc kubenswrapper[4690]: I1203 13:28:46.684885 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" podStartSLOduration=3.803666919 podStartE2EDuration="54.684848871s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.540621247 +0000 UTC m=+1120.521541680" lastFinishedPulling="2025-12-03 13:28:45.421803189 +0000 UTC m=+1171.402723632" observedRunningTime="2025-12-03 13:28:46.675821752 +0000 UTC m=+1172.656742185" watchObservedRunningTime="2025-12-03 13:28:46.684848871 +0000 UTC m=+1172.665769294" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.664394 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" event={"ID":"5f4f6cc8-0d38-4c2a-b4f5-10a870c48479","Type":"ContainerStarted","Data":"f937f7f2750ac671cbe36922c69e34f05887886f7d2289662e8fe57fff468a9c"} Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.666300 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.666454 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" event={"ID":"bc776d95-7342-47e3-b2ad-abc4a3f7aa23","Type":"ContainerStarted","Data":"a9af1996f232f4b9d6b1e294862e4b2e3831c98223e01fe0d90af5063f245007"} Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.666716 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.669679 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" event={"ID":"0e42df6f-0109-4a0b-afeb-630bff5b8c35","Type":"ContainerStarted","Data":"c3b157643b822ff4b365bf6806190d6ce744bd450b251926d14774573b780ea6"} Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.678144 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.700629 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" event={"ID":"c731a6c0-1c65-45e1-893d-f620e83b5363","Type":"ContainerStarted","Data":"54cb2f6772e01ed2c19d4ba85f326392b86b0421f0192c05787dca827159b05c"} Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.706916 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.706974 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" event={"ID":"9890f67c-68ed-475b-9089-262d1fc270b4","Type":"ContainerStarted","Data":"43489a5da1bd5cabf92aa726da7f507ec7ffeb69ca284f5fa9b5a12eaf661dd4"} Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.707584 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" event={"ID":"abb876f0-132b-4224-a3a7-074beba3367b","Type":"ContainerStarted","Data":"b1c204828ff24152b2d3b5e78241a1ad93712c0cb1a8ce39cbb6bff5f7a8ebf1"} Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.708096 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.708347 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.715036 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" event={"ID":"9da3821b-8edf-46c2-ac1c-5cabd4124968","Type":"ContainerStarted","Data":"6ef0fbdcfbf3d8f8bf2ac25bb26656ecced55dd58fbad850e8eddf3678480a38"} Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.715783 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.716113 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.735682 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" podStartSLOduration=3.186004461 podStartE2EDuration="55.735663463s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.444562493 +0000 UTC m=+1120.425482926" lastFinishedPulling="2025-12-03 13:28:46.994221475 +0000 UTC m=+1172.975141928" observedRunningTime="2025-12-03 13:28:47.702285397 +0000 UTC m=+1173.683205830" watchObservedRunningTime="2025-12-03 13:28:47.735663463 +0000 UTC m=+1173.716583896" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.737218 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" podStartSLOduration=3.407098995 podStartE2EDuration="55.737211002s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.539475588 +0000 UTC m=+1120.520396021" lastFinishedPulling="2025-12-03 13:28:46.869587595 +0000 UTC m=+1172.850508028" observedRunningTime="2025-12-03 13:28:47.73123013 +0000 UTC m=+1173.712150563" watchObservedRunningTime="2025-12-03 13:28:47.737211002 +0000 UTC m=+1173.718131435" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.753184 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" podStartSLOduration=3.593146129 podStartE2EDuration="56.753160756s" podCreationTimestamp="2025-12-03 13:27:51 +0000 UTC" firstStartedPulling="2025-12-03 13:27:53.709413874 +0000 UTC m=+1119.690334307" lastFinishedPulling="2025-12-03 13:28:46.869428501 +0000 UTC m=+1172.850348934" observedRunningTime="2025-12-03 13:28:47.75053861 +0000 UTC m=+1173.731459053" watchObservedRunningTime="2025-12-03 13:28:47.753160756 +0000 UTC m=+1173.734081189" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.791665 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" podStartSLOduration=3.337686596 podStartE2EDuration="55.791639152s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.540300269 +0000 UTC m=+1120.521220702" lastFinishedPulling="2025-12-03 13:28:46.994252825 +0000 UTC m=+1172.975173258" observedRunningTime="2025-12-03 13:28:47.774471147 +0000 UTC m=+1173.755391580" watchObservedRunningTime="2025-12-03 13:28:47.791639152 +0000 UTC m=+1173.772559585" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.844381 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" podStartSLOduration=3.147514045 podStartE2EDuration="55.844364949s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.298842888 +0000 UTC m=+1120.279763321" lastFinishedPulling="2025-12-03 13:28:46.995693792 +0000 UTC m=+1172.976614225" observedRunningTime="2025-12-03 13:28:47.842261935 +0000 UTC m=+1173.823182378" watchObservedRunningTime="2025-12-03 13:28:47.844364949 +0000 UTC m=+1173.825285382" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.847692 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" podStartSLOduration=4.248092236 podStartE2EDuration="56.847678783s" podCreationTimestamp="2025-12-03 13:27:51 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.396166236 +0000 UTC m=+1120.377086669" lastFinishedPulling="2025-12-03 13:28:46.995752763 +0000 UTC m=+1172.976673216" observedRunningTime="2025-12-03 13:28:47.820095563 +0000 UTC m=+1173.801016006" watchObservedRunningTime="2025-12-03 13:28:47.847678783 +0000 UTC m=+1173.828599216" Dec 03 13:28:47 crc kubenswrapper[4690]: I1203 13:28:47.886232 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" podStartSLOduration=3.338636941 podStartE2EDuration="55.88621529s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.446703997 +0000 UTC m=+1120.427624430" lastFinishedPulling="2025-12-03 13:28:46.994282346 +0000 UTC m=+1172.975202779" observedRunningTime="2025-12-03 13:28:47.885482371 +0000 UTC m=+1173.866402804" watchObservedRunningTime="2025-12-03 13:28:47.88621529 +0000 UTC m=+1173.867135713" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.249384 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.251829 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-hx22v" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.306670 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnhhq" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.335389 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.343221 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hwxjb" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.348234 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.352114 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-jmz4q" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.420650 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-k6ht5" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.422632 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.428407 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gtcwg" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.528542 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kshv8" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.764693 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-jk4h5" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.769009 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-6g8xn" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.813257 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.840758 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-p559w" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.881234 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5fll5" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.886036 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6q5ss" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.976564 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" Dec 03 13:28:52 crc kubenswrapper[4690]: I1203 13:28:52.980760 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lmq2n" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.028298 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-49f9d" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.173285 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.176807 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7xsmn" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.180589 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.182298 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6xs6t" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.240455 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.242480 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-x7xbr" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.454707 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" Dec 03 13:28:53 crc kubenswrapper[4690]: I1203 13:28:53.457093 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skzwp" Dec 03 13:28:58 crc kubenswrapper[4690]: I1203 13:28:58.380362 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-rrqhz" Dec 03 13:28:58 crc kubenswrapper[4690]: I1203 13:28:58.814464 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd" Dec 03 13:29:03 crc kubenswrapper[4690]: I1203 13:29:03.842122 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" event={"ID":"7df25165-6b23-4dda-8105-05a933bc3ac8","Type":"ContainerStarted","Data":"330042e900fea50a362e0bee5d64b4499aac2f23af175ef542512c123a4f7813"} Dec 03 13:29:03 crc kubenswrapper[4690]: I1203 13:29:03.842901 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" Dec 03 13:29:03 crc kubenswrapper[4690]: I1203 13:29:03.861377 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" podStartSLOduration=3.160527244 podStartE2EDuration="1m11.861357953s" podCreationTimestamp="2025-12-03 13:27:52 +0000 UTC" firstStartedPulling="2025-12-03 13:27:54.694403166 +0000 UTC m=+1120.675323599" lastFinishedPulling="2025-12-03 13:29:03.395233855 +0000 UTC m=+1189.376154308" observedRunningTime="2025-12-03 13:29:03.857421153 +0000 UTC m=+1189.838341586" watchObservedRunningTime="2025-12-03 13:29:03.861357953 +0000 UTC m=+1189.842278376" Dec 03 13:29:13 crc kubenswrapper[4690]: I1203 13:29:13.197958 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d8d6bf864-kzk67" Dec 03 13:29:16 crc kubenswrapper[4690]: I1203 13:29:16.823478 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:29:16 crc kubenswrapper[4690]: I1203 13:29:16.825147 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.497768 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2cm9"] Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.500455 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.509918 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-pkjbq" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.510418 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.510577 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.510795 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.512922 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2cm9"] Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.574965 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7bcm\" (UniqueName: \"kubernetes.io/projected/47594391-e3ad-418f-9054-b17c1a586720-kube-api-access-p7bcm\") pod \"dnsmasq-dns-675f4bcbfc-s2cm9\" (UID: \"47594391-e3ad-418f-9054-b17c1a586720\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.575049 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47594391-e3ad-418f-9054-b17c1a586720-config\") pod \"dnsmasq-dns-675f4bcbfc-s2cm9\" (UID: \"47594391-e3ad-418f-9054-b17c1a586720\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.584614 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2fjw9"] Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.588460 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.591915 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.593502 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2fjw9"] Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.676366 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.676439 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7bcm\" (UniqueName: \"kubernetes.io/projected/47594391-e3ad-418f-9054-b17c1a586720-kube-api-access-p7bcm\") pod \"dnsmasq-dns-675f4bcbfc-s2cm9\" (UID: \"47594391-e3ad-418f-9054-b17c1a586720\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.676487 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47594391-e3ad-418f-9054-b17c1a586720-config\") pod \"dnsmasq-dns-675f4bcbfc-s2cm9\" (UID: \"47594391-e3ad-418f-9054-b17c1a586720\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.676539 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-config\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.676563 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjn4b\" (UniqueName: \"kubernetes.io/projected/d5974090-1232-4acf-a964-ce6c400c5a3b-kube-api-access-pjn4b\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.677360 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47594391-e3ad-418f-9054-b17c1a586720-config\") pod \"dnsmasq-dns-675f4bcbfc-s2cm9\" (UID: \"47594391-e3ad-418f-9054-b17c1a586720\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.695736 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7bcm\" (UniqueName: \"kubernetes.io/projected/47594391-e3ad-418f-9054-b17c1a586720-kube-api-access-p7bcm\") pod \"dnsmasq-dns-675f4bcbfc-s2cm9\" (UID: \"47594391-e3ad-418f-9054-b17c1a586720\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.777829 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.777972 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-config\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.777998 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjn4b\" (UniqueName: \"kubernetes.io/projected/d5974090-1232-4acf-a964-ce6c400c5a3b-kube-api-access-pjn4b\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.778870 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-config\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.779254 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.795135 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjn4b\" (UniqueName: \"kubernetes.io/projected/d5974090-1232-4acf-a964-ce6c400c5a3b-kube-api-access-pjn4b\") pod \"dnsmasq-dns-78dd6ddcc-2fjw9\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.826590 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:29:32 crc kubenswrapper[4690]: I1203 13:29:32.920596 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:29:33 crc kubenswrapper[4690]: I1203 13:29:33.259002 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2cm9"] Dec 03 13:29:33 crc kubenswrapper[4690]: W1203 13:29:33.388992 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5974090_1232_4acf_a964_ce6c400c5a3b.slice/crio-552b7b2371fd7f268f77f29d914f5516f1968add715c77f3fd2d634622de022a WatchSource:0}: Error finding container 552b7b2371fd7f268f77f29d914f5516f1968add715c77f3fd2d634622de022a: Status 404 returned error can't find the container with id 552b7b2371fd7f268f77f29d914f5516f1968add715c77f3fd2d634622de022a Dec 03 13:29:33 crc kubenswrapper[4690]: I1203 13:29:33.390002 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2fjw9"] Dec 03 13:29:34 crc kubenswrapper[4690]: I1203 13:29:34.084660 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" event={"ID":"d5974090-1232-4acf-a964-ce6c400c5a3b","Type":"ContainerStarted","Data":"552b7b2371fd7f268f77f29d914f5516f1968add715c77f3fd2d634622de022a"} Dec 03 13:29:34 crc kubenswrapper[4690]: I1203 13:29:34.085939 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" event={"ID":"47594391-e3ad-418f-9054-b17c1a586720","Type":"ContainerStarted","Data":"78fc94503c5dc26bfe2a737a71e0383c51d77897dd5ba12c86469f8004eb1d39"} Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.032933 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2cm9"] Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.070103 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5mmjs"] Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.071746 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.099142 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5mmjs"] Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.236718 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-config\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.236775 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkbb8\" (UniqueName: \"kubernetes.io/projected/620ef6ec-448c-4457-b82d-0f870a7842ba-kube-api-access-qkbb8\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.236822 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-dns-svc\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.338174 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-config\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.338559 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkbb8\" (UniqueName: \"kubernetes.io/projected/620ef6ec-448c-4457-b82d-0f870a7842ba-kube-api-access-qkbb8\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.338612 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-dns-svc\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.339204 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-config\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.339435 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-dns-svc\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.360746 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkbb8\" (UniqueName: \"kubernetes.io/projected/620ef6ec-448c-4457-b82d-0f870a7842ba-kube-api-access-qkbb8\") pod \"dnsmasq-dns-666b6646f7-5mmjs\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.364519 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2fjw9"] Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.399200 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-glbhd"] Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.402658 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.412107 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.418758 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-glbhd"] Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.543844 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md2rp\" (UniqueName: \"kubernetes.io/projected/b699cbf3-7d22-43cf-8116-52b1de206603-kube-api-access-md2rp\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.543993 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-config\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.544034 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.647731 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md2rp\" (UniqueName: \"kubernetes.io/projected/b699cbf3-7d22-43cf-8116-52b1de206603-kube-api-access-md2rp\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.649835 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-config\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.650051 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.651376 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.652295 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-config\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.672688 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md2rp\" (UniqueName: \"kubernetes.io/projected/b699cbf3-7d22-43cf-8116-52b1de206603-kube-api-access-md2rp\") pod \"dnsmasq-dns-57d769cc4f-glbhd\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:36 crc kubenswrapper[4690]: I1203 13:29:36.740107 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.024549 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5mmjs"] Dec 03 13:29:37 crc kubenswrapper[4690]: W1203 13:29:37.037576 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod620ef6ec_448c_4457_b82d_0f870a7842ba.slice/crio-00104a4be73560b20e041b5ac93401e0690c783348a73a38186079152a10ae07 WatchSource:0}: Error finding container 00104a4be73560b20e041b5ac93401e0690c783348a73a38186079152a10ae07: Status 404 returned error can't find the container with id 00104a4be73560b20e041b5ac93401e0690c783348a73a38186079152a10ae07 Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.135042 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" event={"ID":"620ef6ec-448c-4457-b82d-0f870a7842ba","Type":"ContainerStarted","Data":"00104a4be73560b20e041b5ac93401e0690c783348a73a38186079152a10ae07"} Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.234164 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.236812 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.239584 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.240124 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.240239 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-65kkw" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.240363 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.241406 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.241432 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.247014 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.248094 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.311946 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-glbhd"] Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.376927 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.376981 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pz4h\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-kube-api-access-2pz4h\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377028 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377070 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377096 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377129 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6173a433-90f9-4e48-8297-81767d5ce03e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377163 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6173a433-90f9-4e48-8297-81767d5ce03e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377194 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-config-data\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377218 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377271 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.377303 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.478655 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.478723 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.478760 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.478779 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pz4h\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-kube-api-access-2pz4h\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.478811 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.478851 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.478953 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.478993 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6173a433-90f9-4e48-8297-81767d5ce03e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.479047 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6173a433-90f9-4e48-8297-81767d5ce03e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.479144 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-config-data\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.479174 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.480198 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.480590 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.480700 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-config-data\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.480729 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.481243 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.483203 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.491766 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6173a433-90f9-4e48-8297-81767d5ce03e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.493690 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.494412 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.500669 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6173a433-90f9-4e48-8297-81767d5ce03e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.533006 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pz4h\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-kube-api-access-2pz4h\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.560520 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.562497 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.568174 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-jptwb" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.568338 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.568609 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.568771 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.569030 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.569068 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.569148 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.569941 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.593838 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.682747 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5930069-12cd-4b0b-8aab-0f32ea9057a9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.682811 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.682833 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.682859 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.682938 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.682984 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btxts\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-kube-api-access-btxts\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.683015 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.683043 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.683076 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.683100 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.683139 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5930069-12cd-4b0b-8aab-0f32ea9057a9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.784607 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5930069-12cd-4b0b-8aab-0f32ea9057a9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.784698 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5930069-12cd-4b0b-8aab-0f32ea9057a9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.784728 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.784749 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.784772 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.784794 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.784847 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btxts\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-kube-api-access-btxts\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.785382 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.786330 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.786414 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.786459 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.787354 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.790947 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.792983 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5930069-12cd-4b0b-8aab-0f32ea9057a9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.795037 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.795166 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.795210 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.796570 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5930069-12cd-4b0b-8aab-0f32ea9057a9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.799671 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.804545 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.806392 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btxts\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-kube-api-access-btxts\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.806922 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.829714 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.886181 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:29:37 crc kubenswrapper[4690]: I1203 13:29:37.988329 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.153705 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" event={"ID":"b699cbf3-7d22-43cf-8116-52b1de206603","Type":"ContainerStarted","Data":"a2ea0a928a23df9b2c3b5ff73974263bf1f81bfbc90223708b3bb4752c1c3db6"} Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.489988 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.646444 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.789686 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.791928 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.811002 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.811285 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.811342 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.811508 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-9gc5w" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.813431 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.826185 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.826240 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.826293 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-kolla-config\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.826324 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-config-data-default\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.826362 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5d47d61-0012-4eb8-ba88-da731d54ad45-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.826385 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm2kn\" (UniqueName: \"kubernetes.io/projected/e5d47d61-0012-4eb8-ba88-da731d54ad45-kube-api-access-mm2kn\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.826464 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5d47d61-0012-4eb8-ba88-da731d54ad45-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.826485 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5d47d61-0012-4eb8-ba88-da731d54ad45-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.838938 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.927373 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.927428 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.927460 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-kolla-config\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.927496 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-config-data-default\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.927539 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5d47d61-0012-4eb8-ba88-da731d54ad45-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.927569 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm2kn\" (UniqueName: \"kubernetes.io/projected/e5d47d61-0012-4eb8-ba88-da731d54ad45-kube-api-access-mm2kn\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.927623 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5d47d61-0012-4eb8-ba88-da731d54ad45-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.927651 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5d47d61-0012-4eb8-ba88-da731d54ad45-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.929305 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.929559 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5d47d61-0012-4eb8-ba88-da731d54ad45-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.930390 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-kolla-config\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.932743 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-config-data-default\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.933979 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5d47d61-0012-4eb8-ba88-da731d54ad45-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.956034 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5d47d61-0012-4eb8-ba88-da731d54ad45-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.980348 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm2kn\" (UniqueName: \"kubernetes.io/projected/e5d47d61-0012-4eb8-ba88-da731d54ad45-kube-api-access-mm2kn\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.980860 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5d47d61-0012-4eb8-ba88-da731d54ad45-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:38 crc kubenswrapper[4690]: I1203 13:29:38.988297 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"e5d47d61-0012-4eb8-ba88-da731d54ad45\") " pod="openstack/openstack-galera-0" Dec 03 13:29:39 crc kubenswrapper[4690]: I1203 13:29:39.179241 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.085338 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.088509 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.092359 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.092837 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.093051 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.093183 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-288x9" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.116052 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.270622 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.270701 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.270748 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.270788 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.270927 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c157eb34-b68f-4510-aafd-0c5e73435efe-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.270959 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c157eb34-b68f-4510-aafd-0c5e73435efe-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.270994 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wslkg\" (UniqueName: \"kubernetes.io/projected/c157eb34-b68f-4510-aafd-0c5e73435efe-kube-api-access-wslkg\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.271022 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c157eb34-b68f-4510-aafd-0c5e73435efe-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.372634 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.372678 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.372734 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c157eb34-b68f-4510-aafd-0c5e73435efe-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.372753 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c157eb34-b68f-4510-aafd-0c5e73435efe-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.372777 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wslkg\" (UniqueName: \"kubernetes.io/projected/c157eb34-b68f-4510-aafd-0c5e73435efe-kube-api-access-wslkg\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.372860 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c157eb34-b68f-4510-aafd-0c5e73435efe-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.372942 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.372973 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.373472 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.374871 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.375028 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.375101 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c157eb34-b68f-4510-aafd-0c5e73435efe-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.375105 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c157eb34-b68f-4510-aafd-0c5e73435efe-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.389021 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c157eb34-b68f-4510-aafd-0c5e73435efe-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.389308 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c157eb34-b68f-4510-aafd-0c5e73435efe-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.407312 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wslkg\" (UniqueName: \"kubernetes.io/projected/c157eb34-b68f-4510-aafd-0c5e73435efe-kube-api-access-wslkg\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.472231 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c157eb34-b68f-4510-aafd-0c5e73435efe\") " pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.554592 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.557034 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.561048 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.561135 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.561244 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-7tjhg" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.568410 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.685630 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187a8435-f796-4435-a0b0-211daf924823-combined-ca-bundle\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.685688 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/187a8435-f796-4435-a0b0-211daf924823-config-data\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.685746 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/187a8435-f796-4435-a0b0-211daf924823-memcached-tls-certs\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.685772 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/187a8435-f796-4435-a0b0-211daf924823-kolla-config\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.685793 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb95h\" (UniqueName: \"kubernetes.io/projected/187a8435-f796-4435-a0b0-211daf924823-kube-api-access-hb95h\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.724472 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.787550 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/187a8435-f796-4435-a0b0-211daf924823-memcached-tls-certs\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.787616 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/187a8435-f796-4435-a0b0-211daf924823-kolla-config\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.787645 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb95h\" (UniqueName: \"kubernetes.io/projected/187a8435-f796-4435-a0b0-211daf924823-kube-api-access-hb95h\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.787732 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187a8435-f796-4435-a0b0-211daf924823-combined-ca-bundle\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.787759 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/187a8435-f796-4435-a0b0-211daf924823-config-data\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.788740 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/187a8435-f796-4435-a0b0-211daf924823-kolla-config\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.789219 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/187a8435-f796-4435-a0b0-211daf924823-config-data\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.795015 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/187a8435-f796-4435-a0b0-211daf924823-memcached-tls-certs\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.797504 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187a8435-f796-4435-a0b0-211daf924823-combined-ca-bundle\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.809510 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb95h\" (UniqueName: \"kubernetes.io/projected/187a8435-f796-4435-a0b0-211daf924823-kube-api-access-hb95h\") pod \"memcached-0\" (UID: \"187a8435-f796-4435-a0b0-211daf924823\") " pod="openstack/memcached-0" Dec 03 13:29:40 crc kubenswrapper[4690]: I1203 13:29:40.907648 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 13:29:43 crc kubenswrapper[4690]: I1203 13:29:43.046691 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:29:43 crc kubenswrapper[4690]: I1203 13:29:43.048314 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:29:43 crc kubenswrapper[4690]: I1203 13:29:43.056915 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:29:43 crc kubenswrapper[4690]: I1203 13:29:43.060426 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-hzk77" Dec 03 13:29:43 crc kubenswrapper[4690]: I1203 13:29:43.136431 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5mlv\" (UniqueName: \"kubernetes.io/projected/7b78850b-78ce-4668-97e8-d4929857b34a-kube-api-access-s5mlv\") pod \"kube-state-metrics-0\" (UID: \"7b78850b-78ce-4668-97e8-d4929857b34a\") " pod="openstack/kube-state-metrics-0" Dec 03 13:29:43 crc kubenswrapper[4690]: I1203 13:29:43.238794 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5mlv\" (UniqueName: \"kubernetes.io/projected/7b78850b-78ce-4668-97e8-d4929857b34a-kube-api-access-s5mlv\") pod \"kube-state-metrics-0\" (UID: \"7b78850b-78ce-4668-97e8-d4929857b34a\") " pod="openstack/kube-state-metrics-0" Dec 03 13:29:43 crc kubenswrapper[4690]: I1203 13:29:43.271677 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5mlv\" (UniqueName: \"kubernetes.io/projected/7b78850b-78ce-4668-97e8-d4929857b34a-kube-api-access-s5mlv\") pod \"kube-state-metrics-0\" (UID: \"7b78850b-78ce-4668-97e8-d4929857b34a\") " pod="openstack/kube-state-metrics-0" Dec 03 13:29:43 crc kubenswrapper[4690]: I1203 13:29:43.378544 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.223533 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6173a433-90f9-4e48-8297-81767d5ce03e","Type":"ContainerStarted","Data":"cc86a2e8021c328d4a576e674888abb0efa60b2e6200e716205d5af605ecd167"} Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.224867 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5930069-12cd-4b0b-8aab-0f32ea9057a9","Type":"ContainerStarted","Data":"8af360905b9e2c3a677ccd5b8c32087eb23ee9c42a552ff718e90680f92f716c"} Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.378372 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.380371 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.385288 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.386199 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.386858 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-7zt5s" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.388662 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.390595 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.393378 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.511239 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm8zj\" (UniqueName: \"kubernetes.io/projected/c44ebdaa-211b-4ac6-90c6-5236eaee6295-kube-api-access-pm8zj\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.511310 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c44ebdaa-211b-4ac6-90c6-5236eaee6295-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.511361 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.511521 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c44ebdaa-211b-4ac6-90c6-5236eaee6295-config\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.511628 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c44ebdaa-211b-4ac6-90c6-5236eaee6295-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.511729 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.511929 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.512069 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.614765 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.614925 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm8zj\" (UniqueName: \"kubernetes.io/projected/c44ebdaa-211b-4ac6-90c6-5236eaee6295-kube-api-access-pm8zj\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.614971 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c44ebdaa-211b-4ac6-90c6-5236eaee6295-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.615210 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.615408 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.615457 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c44ebdaa-211b-4ac6-90c6-5236eaee6295-config\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.615518 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c44ebdaa-211b-4ac6-90c6-5236eaee6295-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.615674 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.615805 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.616096 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c44ebdaa-211b-4ac6-90c6-5236eaee6295-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.616807 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c44ebdaa-211b-4ac6-90c6-5236eaee6295-config\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.616946 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c44ebdaa-211b-4ac6-90c6-5236eaee6295-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.625266 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.625410 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.629903 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c44ebdaa-211b-4ac6-90c6-5236eaee6295-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.640381 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm8zj\" (UniqueName: \"kubernetes.io/projected/c44ebdaa-211b-4ac6-90c6-5236eaee6295-kube-api-access-pm8zj\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.644855 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c44ebdaa-211b-4ac6-90c6-5236eaee6295\") " pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.760122 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.823957 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.824059 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.996783 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mf8xw"] Dec 03 13:29:46 crc kubenswrapper[4690]: I1203 13:29:46.998899 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.001021 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-rksz7" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.001605 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.004258 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.037532 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-znnxj"] Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.050592 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.069816 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mf8xw"] Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.103017 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-znnxj"] Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150408 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-log\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150470 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-etc-ovs\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150522 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-scripts\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150544 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-log-ovn\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150623 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/64710eed-cdda-4f62-a007-766dd92f0b29-ovn-controller-tls-certs\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150661 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-lib\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150689 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-run\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150718 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6prd7\" (UniqueName: \"kubernetes.io/projected/64710eed-cdda-4f62-a007-766dd92f0b29-kube-api-access-6prd7\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150746 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64710eed-cdda-4f62-a007-766dd92f0b29-combined-ca-bundle\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150779 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64710eed-cdda-4f62-a007-766dd92f0b29-scripts\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150804 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbzjz\" (UniqueName: \"kubernetes.io/projected/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-kube-api-access-jbzjz\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150840 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-run-ovn\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.150893 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-run\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.252725 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-log\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.252787 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-etc-ovs\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.252812 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-scripts\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.252835 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-log-ovn\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.252912 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/64710eed-cdda-4f62-a007-766dd92f0b29-ovn-controller-tls-certs\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.252942 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-lib\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.252966 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-run\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.252994 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6prd7\" (UniqueName: \"kubernetes.io/projected/64710eed-cdda-4f62-a007-766dd92f0b29-kube-api-access-6prd7\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253019 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64710eed-cdda-4f62-a007-766dd92f0b29-combined-ca-bundle\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253053 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64710eed-cdda-4f62-a007-766dd92f0b29-scripts\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253072 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbzjz\" (UniqueName: \"kubernetes.io/projected/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-kube-api-access-jbzjz\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253108 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-run-ovn\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253139 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-run\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253562 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-log\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253641 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-run\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253690 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-run\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.253942 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-etc-ovs\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.255036 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-run-ovn\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.255217 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/64710eed-cdda-4f62-a007-766dd92f0b29-var-log-ovn\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.255286 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-var-lib\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.256672 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-scripts\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.257692 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64710eed-cdda-4f62-a007-766dd92f0b29-scripts\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.273742 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/64710eed-cdda-4f62-a007-766dd92f0b29-ovn-controller-tls-certs\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.273753 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64710eed-cdda-4f62-a007-766dd92f0b29-combined-ca-bundle\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.276313 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbzjz\" (UniqueName: \"kubernetes.io/projected/cf553f8a-b61b-469b-9c0c-d602d73ef4c2-kube-api-access-jbzjz\") pod \"ovn-controller-ovs-znnxj\" (UID: \"cf553f8a-b61b-469b-9c0c-d602d73ef4c2\") " pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.278373 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6prd7\" (UniqueName: \"kubernetes.io/projected/64710eed-cdda-4f62-a007-766dd92f0b29-kube-api-access-6prd7\") pod \"ovn-controller-mf8xw\" (UID: \"64710eed-cdda-4f62-a007-766dd92f0b29\") " pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.352233 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mf8xw" Dec 03 13:29:47 crc kubenswrapper[4690]: I1203 13:29:47.376131 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:29:49 crc kubenswrapper[4690]: I1203 13:29:49.176146 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.441904 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.451293 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.454313 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.454635 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.454814 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.455004 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-9lbdl" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.456646 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.530561 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94pvv\" (UniqueName: \"kubernetes.io/projected/d19eec48-0864-43d8-8071-1aef21a26286-kube-api-access-94pvv\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.530674 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d19eec48-0864-43d8-8071-1aef21a26286-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.530724 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d19eec48-0864-43d8-8071-1aef21a26286-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.530745 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.530778 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.530813 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19eec48-0864-43d8-8071-1aef21a26286-config\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.530843 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.530922 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634178 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d19eec48-0864-43d8-8071-1aef21a26286-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634243 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634295 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634361 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19eec48-0864-43d8-8071-1aef21a26286-config\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634394 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634457 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634548 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94pvv\" (UniqueName: \"kubernetes.io/projected/d19eec48-0864-43d8-8071-1aef21a26286-kube-api-access-94pvv\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634613 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d19eec48-0864-43d8-8071-1aef21a26286-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.634924 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.635137 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d19eec48-0864-43d8-8071-1aef21a26286-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.635460 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d19eec48-0864-43d8-8071-1aef21a26286-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.635560 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19eec48-0864-43d8-8071-1aef21a26286-config\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.640925 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.642624 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.668969 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19eec48-0864-43d8-8071-1aef21a26286-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.676942 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94pvv\" (UniqueName: \"kubernetes.io/projected/d19eec48-0864-43d8-8071-1aef21a26286-kube-api-access-94pvv\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.713240 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d19eec48-0864-43d8-8071-1aef21a26286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 13:29:50 crc kubenswrapper[4690]: I1203 13:29:50.782528 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.154181 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2"] Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.156190 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.159029 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.159403 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.161726 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2"] Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.308928 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8wjz\" (UniqueName: \"kubernetes.io/projected/d9edbc53-d827-45be-aa22-bb691422fb4d-kube-api-access-r8wjz\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.309211 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9edbc53-d827-45be-aa22-bb691422fb4d-config-volume\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.309345 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9edbc53-d827-45be-aa22-bb691422fb4d-secret-volume\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.411577 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9edbc53-d827-45be-aa22-bb691422fb4d-secret-volume\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.411668 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8wjz\" (UniqueName: \"kubernetes.io/projected/d9edbc53-d827-45be-aa22-bb691422fb4d-kube-api-access-r8wjz\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.411762 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9edbc53-d827-45be-aa22-bb691422fb4d-config-volume\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.412997 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9edbc53-d827-45be-aa22-bb691422fb4d-config-volume\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.420547 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9edbc53-d827-45be-aa22-bb691422fb4d-secret-volume\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.439388 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8wjz\" (UniqueName: \"kubernetes.io/projected/d9edbc53-d827-45be-aa22-bb691422fb4d-kube-api-access-r8wjz\") pod \"collect-profiles-29412810-x2cl2\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:00 crc kubenswrapper[4690]: I1203 13:30:00.489914 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:03 crc kubenswrapper[4690]: I1203 13:30:03.383689 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 13:30:03 crc kubenswrapper[4690]: I1203 13:30:03.391325 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c157eb34-b68f-4510-aafd-0c5e73435efe","Type":"ContainerStarted","Data":"0a1e0c7ec49489dbc1797ae2102d32f6e04d0a143d022e9f4ba217d4e71012cc"} Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.394255 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.395387 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pjn4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-2fjw9_openstack(d5974090-1232-4acf-a964-ce6c400c5a3b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.397094 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" podUID="d5974090-1232-4acf-a964-ce6c400c5a3b" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.415163 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.415353 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p7bcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-s2cm9_openstack(47594391-e3ad-418f-9054-b17c1a586720): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.416595 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" podUID="47594391-e3ad-418f-9054-b17c1a586720" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.630596 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.630822 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qkbb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-5mmjs_openstack(620ef6ec-448c-4457-b82d-0f870a7842ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.632319 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" podUID="620ef6ec-448c-4457-b82d-0f870a7842ba" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.755592 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.755810 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-md2rp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-glbhd_openstack(b699cbf3-7d22-43cf-8116-52b1de206603): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:30:06 crc kubenswrapper[4690]: E1203 13:30:06.757299 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" podUID="b699cbf3-7d22-43cf-8116-52b1de206603" Dec 03 13:30:07 crc kubenswrapper[4690]: E1203 13:30:07.434308 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" podUID="b699cbf3-7d22-43cf-8116-52b1de206603" Dec 03 13:30:07 crc kubenswrapper[4690]: E1203 13:30:07.434313 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" podUID="620ef6ec-448c-4457-b82d-0f870a7842ba" Dec 03 13:30:07 crc kubenswrapper[4690]: W1203 13:30:07.917309 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod187a8435_f796_4435_a0b0_211daf924823.slice/crio-566a07bf81c4abde80473f1b5d24210aa6ec22bf6fb56e40a2eb069291404c30 WatchSource:0}: Error finding container 566a07bf81c4abde80473f1b5d24210aa6ec22bf6fb56e40a2eb069291404c30: Status 404 returned error can't find the container with id 566a07bf81c4abde80473f1b5d24210aa6ec22bf6fb56e40a2eb069291404c30 Dec 03 13:30:08 crc kubenswrapper[4690]: E1203 13:30:08.072676 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 13:30:08 crc kubenswrapper[4690]: E1203 13:30:08.073216 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-btxts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(d5930069-12cd-4b0b-8aab-0f32ea9057a9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:30:08 crc kubenswrapper[4690]: E1203 13:30:08.074598 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.127915 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.169452 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:30:08 crc kubenswrapper[4690]: E1203 13:30:08.173113 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 13:30:08 crc kubenswrapper[4690]: E1203 13:30:08.173393 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2pz4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(6173a433-90f9-4e48-8297-81767d5ce03e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:30:08 crc kubenswrapper[4690]: E1203 13:30:08.174652 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.304857 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7bcm\" (UniqueName: \"kubernetes.io/projected/47594391-e3ad-418f-9054-b17c1a586720-kube-api-access-p7bcm\") pod \"47594391-e3ad-418f-9054-b17c1a586720\" (UID: \"47594391-e3ad-418f-9054-b17c1a586720\") " Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.305077 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjn4b\" (UniqueName: \"kubernetes.io/projected/d5974090-1232-4acf-a964-ce6c400c5a3b-kube-api-access-pjn4b\") pod \"d5974090-1232-4acf-a964-ce6c400c5a3b\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.305165 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-config\") pod \"d5974090-1232-4acf-a964-ce6c400c5a3b\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.305194 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-dns-svc\") pod \"d5974090-1232-4acf-a964-ce6c400c5a3b\" (UID: \"d5974090-1232-4acf-a964-ce6c400c5a3b\") " Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.305215 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47594391-e3ad-418f-9054-b17c1a586720-config\") pod \"47594391-e3ad-418f-9054-b17c1a586720\" (UID: \"47594391-e3ad-418f-9054-b17c1a586720\") " Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.305706 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-config" (OuterVolumeSpecName: "config") pod "d5974090-1232-4acf-a964-ce6c400c5a3b" (UID: "d5974090-1232-4acf-a964-ce6c400c5a3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.305757 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5974090-1232-4acf-a964-ce6c400c5a3b" (UID: "d5974090-1232-4acf-a964-ce6c400c5a3b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.305849 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47594391-e3ad-418f-9054-b17c1a586720-config" (OuterVolumeSpecName: "config") pod "47594391-e3ad-418f-9054-b17c1a586720" (UID: "47594391-e3ad-418f-9054-b17c1a586720"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.312331 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47594391-e3ad-418f-9054-b17c1a586720-kube-api-access-p7bcm" (OuterVolumeSpecName: "kube-api-access-p7bcm") pod "47594391-e3ad-418f-9054-b17c1a586720" (UID: "47594391-e3ad-418f-9054-b17c1a586720"). InnerVolumeSpecName "kube-api-access-p7bcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.319858 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5974090-1232-4acf-a964-ce6c400c5a3b-kube-api-access-pjn4b" (OuterVolumeSpecName: "kube-api-access-pjn4b") pod "d5974090-1232-4acf-a964-ce6c400c5a3b" (UID: "d5974090-1232-4acf-a964-ce6c400c5a3b"). InnerVolumeSpecName "kube-api-access-pjn4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.407292 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7bcm\" (UniqueName: \"kubernetes.io/projected/47594391-e3ad-418f-9054-b17c1a586720-kube-api-access-p7bcm\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.407334 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjn4b\" (UniqueName: \"kubernetes.io/projected/d5974090-1232-4acf-a964-ce6c400c5a3b-kube-api-access-pjn4b\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.407348 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.407359 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5974090-1232-4acf-a964-ce6c400c5a3b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.407367 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47594391-e3ad-418f-9054-b17c1a586720-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.437510 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" event={"ID":"d5974090-1232-4acf-a964-ce6c400c5a3b","Type":"ContainerDied","Data":"552b7b2371fd7f268f77f29d914f5516f1968add715c77f3fd2d634622de022a"} Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.437607 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2fjw9" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.440258 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" event={"ID":"47594391-e3ad-418f-9054-b17c1a586720","Type":"ContainerDied","Data":"78fc94503c5dc26bfe2a737a71e0383c51d77897dd5ba12c86469f8004eb1d39"} Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.440320 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s2cm9" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.449251 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"187a8435-f796-4435-a0b0-211daf924823","Type":"ContainerStarted","Data":"566a07bf81c4abde80473f1b5d24210aa6ec22bf6fb56e40a2eb069291404c30"} Dec 03 13:30:08 crc kubenswrapper[4690]: E1203 13:30:08.451505 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" Dec 03 13:30:08 crc kubenswrapper[4690]: E1203 13:30:08.451827 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.495120 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2fjw9"] Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.517581 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2fjw9"] Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.562389 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2cm9"] Dec 03 13:30:08 crc kubenswrapper[4690]: W1203 13:30:08.564853 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5d47d61_0012_4eb8_ba88_da731d54ad45.slice/crio-19a493bc7711c558dd2003fc4ad36262e0da3e54f61fc07a585bbfb8c181a66e WatchSource:0}: Error finding container 19a493bc7711c558dd2003fc4ad36262e0da3e54f61fc07a585bbfb8c181a66e: Status 404 returned error can't find the container with id 19a493bc7711c558dd2003fc4ad36262e0da3e54f61fc07a585bbfb8c181a66e Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.571791 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2cm9"] Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.588241 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.684034 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.716413 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:30:08 crc kubenswrapper[4690]: W1203 13:30:08.725509 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b78850b_78ce_4668_97e8_d4929857b34a.slice/crio-c122fbdbfa57364abcec5cbe90ef1d77c4cb7e66f5a94559a4c3b28c27cc8f50 WatchSource:0}: Error finding container c122fbdbfa57364abcec5cbe90ef1d77c4cb7e66f5a94559a4c3b28c27cc8f50: Status 404 returned error can't find the container with id c122fbdbfa57364abcec5cbe90ef1d77c4cb7e66f5a94559a4c3b28c27cc8f50 Dec 03 13:30:08 crc kubenswrapper[4690]: W1203 13:30:08.731850 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9edbc53_d827_45be_aa22_bb691422fb4d.slice/crio-a35c188dc1ba0b2a7acd29dee175006759fc455ebbd9838608bbf08edf679713 WatchSource:0}: Error finding container a35c188dc1ba0b2a7acd29dee175006759fc455ebbd9838608bbf08edf679713: Status 404 returned error can't find the container with id a35c188dc1ba0b2a7acd29dee175006759fc455ebbd9838608bbf08edf679713 Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.792100 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2"] Dec 03 13:30:08 crc kubenswrapper[4690]: W1203 13:30:08.840764 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64710eed_cdda_4f62_a007_766dd92f0b29.slice/crio-5525a0b866a30f9a8871e0472961adb3f3f7f02d7b78426f52c4728a1ffa90a2 WatchSource:0}: Error finding container 5525a0b866a30f9a8871e0472961adb3f3f7f02d7b78426f52c4728a1ffa90a2: Status 404 returned error can't find the container with id 5525a0b866a30f9a8871e0472961adb3f3f7f02d7b78426f52c4728a1ffa90a2 Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.844108 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mf8xw"] Dec 03 13:30:08 crc kubenswrapper[4690]: I1203 13:30:08.910401 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.030943 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-znnxj"] Dec 03 13:30:09 crc kubenswrapper[4690]: W1203 13:30:09.040492 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf553f8a_b61b_469b_9c0c_d602d73ef4c2.slice/crio-95b603d9a33f351794f2b544ff1e520d787328fb28dcc226c0cc3a56ffa6ed6f WatchSource:0}: Error finding container 95b603d9a33f351794f2b544ff1e520d787328fb28dcc226c0cc3a56ffa6ed6f: Status 404 returned error can't find the container with id 95b603d9a33f351794f2b544ff1e520d787328fb28dcc226c0cc3a56ffa6ed6f Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.460444 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7b78850b-78ce-4668-97e8-d4929857b34a","Type":"ContainerStarted","Data":"c122fbdbfa57364abcec5cbe90ef1d77c4cb7e66f5a94559a4c3b28c27cc8f50"} Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.463258 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c44ebdaa-211b-4ac6-90c6-5236eaee6295","Type":"ContainerStarted","Data":"97b15d622e7508b9955790d83eb2c886c600c77ceb5f12a2b791445470a2cabb"} Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.466436 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mf8xw" event={"ID":"64710eed-cdda-4f62-a007-766dd92f0b29","Type":"ContainerStarted","Data":"5525a0b866a30f9a8871e0472961adb3f3f7f02d7b78426f52c4728a1ffa90a2"} Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.468501 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-znnxj" event={"ID":"cf553f8a-b61b-469b-9c0c-d602d73ef4c2","Type":"ContainerStarted","Data":"95b603d9a33f351794f2b544ff1e520d787328fb28dcc226c0cc3a56ffa6ed6f"} Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.471596 4690 generic.go:334] "Generic (PLEG): container finished" podID="d9edbc53-d827-45be-aa22-bb691422fb4d" containerID="227845436ebc018f13d5937233cfc10d0acbf689be325a7635063e7be03af4cd" exitCode=0 Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.471663 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" event={"ID":"d9edbc53-d827-45be-aa22-bb691422fb4d","Type":"ContainerDied","Data":"227845436ebc018f13d5937233cfc10d0acbf689be325a7635063e7be03af4cd"} Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.471680 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" event={"ID":"d9edbc53-d827-45be-aa22-bb691422fb4d","Type":"ContainerStarted","Data":"a35c188dc1ba0b2a7acd29dee175006759fc455ebbd9838608bbf08edf679713"} Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.473614 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d19eec48-0864-43d8-8071-1aef21a26286","Type":"ContainerStarted","Data":"b7b39e97cb14648c944592449a2a9a9c54504441eaaf863d82cc16f4e23af005"} Dec 03 13:30:09 crc kubenswrapper[4690]: I1203 13:30:09.477017 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e5d47d61-0012-4eb8-ba88-da731d54ad45","Type":"ContainerStarted","Data":"19a493bc7711c558dd2003fc4ad36262e0da3e54f61fc07a585bbfb8c181a66e"} Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.329808 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47594391-e3ad-418f-9054-b17c1a586720" path="/var/lib/kubelet/pods/47594391-e3ad-418f-9054-b17c1a586720/volumes" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.331201 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5974090-1232-4acf-a964-ce6c400c5a3b" path="/var/lib/kubelet/pods/d5974090-1232-4acf-a964-ce6c400c5a3b/volumes" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.473794 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-wjlxx"] Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.475597 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.478518 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.482525 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-wjlxx"] Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.573511 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qqlk\" (UniqueName: \"kubernetes.io/projected/338580e8-66c9-4169-93f5-1e7f530de31c-kube-api-access-9qqlk\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.573587 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338580e8-66c9-4169-93f5-1e7f530de31c-combined-ca-bundle\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.573658 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/338580e8-66c9-4169-93f5-1e7f530de31c-ovs-rundir\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.573701 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/338580e8-66c9-4169-93f5-1e7f530de31c-config\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.573767 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/338580e8-66c9-4169-93f5-1e7f530de31c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.573808 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/338580e8-66c9-4169-93f5-1e7f530de31c-ovn-rundir\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.676080 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/338580e8-66c9-4169-93f5-1e7f530de31c-ovs-rundir\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.676155 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/338580e8-66c9-4169-93f5-1e7f530de31c-config\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.676265 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/338580e8-66c9-4169-93f5-1e7f530de31c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.676326 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/338580e8-66c9-4169-93f5-1e7f530de31c-ovn-rundir\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.676368 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qqlk\" (UniqueName: \"kubernetes.io/projected/338580e8-66c9-4169-93f5-1e7f530de31c-kube-api-access-9qqlk\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.676434 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338580e8-66c9-4169-93f5-1e7f530de31c-combined-ca-bundle\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.680336 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/338580e8-66c9-4169-93f5-1e7f530de31c-ovn-rundir\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.680450 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/338580e8-66c9-4169-93f5-1e7f530de31c-ovs-rundir\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.681357 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/338580e8-66c9-4169-93f5-1e7f530de31c-config\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.688996 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338580e8-66c9-4169-93f5-1e7f530de31c-combined-ca-bundle\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.699089 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/338580e8-66c9-4169-93f5-1e7f530de31c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.719714 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qqlk\" (UniqueName: \"kubernetes.io/projected/338580e8-66c9-4169-93f5-1e7f530de31c-kube-api-access-9qqlk\") pod \"ovn-controller-metrics-wjlxx\" (UID: \"338580e8-66c9-4169-93f5-1e7f530de31c\") " pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.746675 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5mmjs"] Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.797391 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-wjlxx" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.851811 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-h2hnf"] Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.853400 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.871458 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.903570 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-h2hnf"] Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.987532 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.987664 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.987700 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kpxw\" (UniqueName: \"kubernetes.io/projected/09ac08bf-8841-48a1-94ba-0541e5596c5c-kube-api-access-2kpxw\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:10 crc kubenswrapper[4690]: I1203 13:30:10.987759 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-config\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.089674 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-config\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.089764 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.089833 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.089850 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kpxw\" (UniqueName: \"kubernetes.io/projected/09ac08bf-8841-48a1-94ba-0541e5596c5c-kube-api-access-2kpxw\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.091014 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-config\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.091551 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.092427 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.132438 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kpxw\" (UniqueName: \"kubernetes.io/projected/09ac08bf-8841-48a1-94ba-0541e5596c5c-kube-api-access-2kpxw\") pod \"dnsmasq-dns-7fd796d7df-h2hnf\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.149696 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-glbhd"] Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.185269 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kvkvr"] Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.186773 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.194551 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.215327 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.216233 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kvkvr"] Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.296335 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl9x9\" (UniqueName: \"kubernetes.io/projected/4ce1db85-6b25-47fd-84ee-47406aa77eec-kube-api-access-hl9x9\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.296732 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.296779 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-config\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.296837 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.296856 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.310127 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.364282 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.428349 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9edbc53-d827-45be-aa22-bb691422fb4d-secret-volume\") pod \"d9edbc53-d827-45be-aa22-bb691422fb4d\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.428436 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8wjz\" (UniqueName: \"kubernetes.io/projected/d9edbc53-d827-45be-aa22-bb691422fb4d-kube-api-access-r8wjz\") pod \"d9edbc53-d827-45be-aa22-bb691422fb4d\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.428723 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9edbc53-d827-45be-aa22-bb691422fb4d-config-volume\") pod \"d9edbc53-d827-45be-aa22-bb691422fb4d\" (UID: \"d9edbc53-d827-45be-aa22-bb691422fb4d\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.429559 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.429677 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.429959 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl9x9\" (UniqueName: \"kubernetes.io/projected/4ce1db85-6b25-47fd-84ee-47406aa77eec-kube-api-access-hl9x9\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.430133 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.430293 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-config\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.430354 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9edbc53-d827-45be-aa22-bb691422fb4d-config-volume" (OuterVolumeSpecName: "config-volume") pod "d9edbc53-d827-45be-aa22-bb691422fb4d" (UID: "d9edbc53-d827-45be-aa22-bb691422fb4d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.430659 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.430860 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.431328 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-config\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.431650 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.453995 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9edbc53-d827-45be-aa22-bb691422fb4d-kube-api-access-r8wjz" (OuterVolumeSpecName: "kube-api-access-r8wjz") pod "d9edbc53-d827-45be-aa22-bb691422fb4d" (UID: "d9edbc53-d827-45be-aa22-bb691422fb4d"). InnerVolumeSpecName "kube-api-access-r8wjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.457231 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9edbc53-d827-45be-aa22-bb691422fb4d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d9edbc53-d827-45be-aa22-bb691422fb4d" (UID: "d9edbc53-d827-45be-aa22-bb691422fb4d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.463025 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl9x9\" (UniqueName: \"kubernetes.io/projected/4ce1db85-6b25-47fd-84ee-47406aa77eec-kube-api-access-hl9x9\") pod \"dnsmasq-dns-86db49b7ff-kvkvr\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.507313 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" event={"ID":"d9edbc53-d827-45be-aa22-bb691422fb4d","Type":"ContainerDied","Data":"a35c188dc1ba0b2a7acd29dee175006759fc455ebbd9838608bbf08edf679713"} Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.507372 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a35c188dc1ba0b2a7acd29dee175006759fc455ebbd9838608bbf08edf679713" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.507462 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.517357 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" event={"ID":"620ef6ec-448c-4457-b82d-0f870a7842ba","Type":"ContainerDied","Data":"00104a4be73560b20e041b5ac93401e0690c783348a73a38186079152a10ae07"} Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.517494 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-5mmjs" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.531510 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-config\") pod \"620ef6ec-448c-4457-b82d-0f870a7842ba\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.531972 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-dns-svc\") pod \"620ef6ec-448c-4457-b82d-0f870a7842ba\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.532041 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkbb8\" (UniqueName: \"kubernetes.io/projected/620ef6ec-448c-4457-b82d-0f870a7842ba-kube-api-access-qkbb8\") pod \"620ef6ec-448c-4457-b82d-0f870a7842ba\" (UID: \"620ef6ec-448c-4457-b82d-0f870a7842ba\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.532446 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9edbc53-d827-45be-aa22-bb691422fb4d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.532459 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8wjz\" (UniqueName: \"kubernetes.io/projected/d9edbc53-d827-45be-aa22-bb691422fb4d-kube-api-access-r8wjz\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.532467 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9edbc53-d827-45be-aa22-bb691422fb4d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.533372 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-config" (OuterVolumeSpecName: "config") pod "620ef6ec-448c-4457-b82d-0f870a7842ba" (UID: "620ef6ec-448c-4457-b82d-0f870a7842ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.533620 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "620ef6ec-448c-4457-b82d-0f870a7842ba" (UID: "620ef6ec-448c-4457-b82d-0f870a7842ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.535614 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/620ef6ec-448c-4457-b82d-0f870a7842ba-kube-api-access-qkbb8" (OuterVolumeSpecName: "kube-api-access-qkbb8") pod "620ef6ec-448c-4457-b82d-0f870a7842ba" (UID: "620ef6ec-448c-4457-b82d-0f870a7842ba"). InnerVolumeSpecName "kube-api-access-qkbb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.541555 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.585539 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.634223 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-dns-svc\") pod \"b699cbf3-7d22-43cf-8116-52b1de206603\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.634366 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-config\") pod \"b699cbf3-7d22-43cf-8116-52b1de206603\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.634444 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md2rp\" (UniqueName: \"kubernetes.io/projected/b699cbf3-7d22-43cf-8116-52b1de206603-kube-api-access-md2rp\") pod \"b699cbf3-7d22-43cf-8116-52b1de206603\" (UID: \"b699cbf3-7d22-43cf-8116-52b1de206603\") " Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.634805 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b699cbf3-7d22-43cf-8116-52b1de206603" (UID: "b699cbf3-7d22-43cf-8116-52b1de206603"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.634843 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-config" (OuterVolumeSpecName: "config") pod "b699cbf3-7d22-43cf-8116-52b1de206603" (UID: "b699cbf3-7d22-43cf-8116-52b1de206603"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.635654 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.635672 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b699cbf3-7d22-43cf-8116-52b1de206603-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.635683 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.635727 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkbb8\" (UniqueName: \"kubernetes.io/projected/620ef6ec-448c-4457-b82d-0f870a7842ba-kube-api-access-qkbb8\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.635747 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/620ef6ec-448c-4457-b82d-0f870a7842ba-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.640685 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b699cbf3-7d22-43cf-8116-52b1de206603-kube-api-access-md2rp" (OuterVolumeSpecName: "kube-api-access-md2rp") pod "b699cbf3-7d22-43cf-8116-52b1de206603" (UID: "b699cbf3-7d22-43cf-8116-52b1de206603"). InnerVolumeSpecName "kube-api-access-md2rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.737293 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md2rp\" (UniqueName: \"kubernetes.io/projected/b699cbf3-7d22-43cf-8116-52b1de206603-kube-api-access-md2rp\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.805505 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-wjlxx"] Dec 03 13:30:11 crc kubenswrapper[4690]: W1203 13:30:11.813668 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod338580e8_66c9_4169_93f5_1e7f530de31c.slice/crio-0eda41df066ed831f82f0af84ef0df5e56fc471ef76fb9c864a2a3c973c6f601 WatchSource:0}: Error finding container 0eda41df066ed831f82f0af84ef0df5e56fc471ef76fb9c864a2a3c973c6f601: Status 404 returned error can't find the container with id 0eda41df066ed831f82f0af84ef0df5e56fc471ef76fb9c864a2a3c973c6f601 Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.904007 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5mmjs"] Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.920110 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5mmjs"] Dec 03 13:30:11 crc kubenswrapper[4690]: I1203 13:30:11.930610 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-h2hnf"] Dec 03 13:30:12 crc kubenswrapper[4690]: I1203 13:30:12.088798 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kvkvr"] Dec 03 13:30:12 crc kubenswrapper[4690]: I1203 13:30:12.328608 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="620ef6ec-448c-4457-b82d-0f870a7842ba" path="/var/lib/kubelet/pods/620ef6ec-448c-4457-b82d-0f870a7842ba/volumes" Dec 03 13:30:12 crc kubenswrapper[4690]: I1203 13:30:12.529807 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-wjlxx" event={"ID":"338580e8-66c9-4169-93f5-1e7f530de31c","Type":"ContainerStarted","Data":"0eda41df066ed831f82f0af84ef0df5e56fc471ef76fb9c864a2a3c973c6f601"} Dec 03 13:30:12 crc kubenswrapper[4690]: I1203 13:30:12.532111 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" event={"ID":"b699cbf3-7d22-43cf-8116-52b1de206603","Type":"ContainerDied","Data":"a2ea0a928a23df9b2c3b5ff73974263bf1f81bfbc90223708b3bb4752c1c3db6"} Dec 03 13:30:12 crc kubenswrapper[4690]: I1203 13:30:12.532249 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-glbhd" Dec 03 13:30:12 crc kubenswrapper[4690]: I1203 13:30:12.578919 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-glbhd"] Dec 03 13:30:12 crc kubenswrapper[4690]: I1203 13:30:12.588029 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-glbhd"] Dec 03 13:30:14 crc kubenswrapper[4690]: I1203 13:30:14.332380 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b699cbf3-7d22-43cf-8116-52b1de206603" path="/var/lib/kubelet/pods/b699cbf3-7d22-43cf-8116-52b1de206603/volumes" Dec 03 13:30:15 crc kubenswrapper[4690]: W1203 13:30:15.957137 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09ac08bf_8841_48a1_94ba_0541e5596c5c.slice/crio-dd5f779a44e0378bbef8e71d300541727499a7027732a574daddc97550b70c99 WatchSource:0}: Error finding container dd5f779a44e0378bbef8e71d300541727499a7027732a574daddc97550b70c99: Status 404 returned error can't find the container with id dd5f779a44e0378bbef8e71d300541727499a7027732a574daddc97550b70c99 Dec 03 13:30:16 crc kubenswrapper[4690]: I1203 13:30:16.585475 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" event={"ID":"09ac08bf-8841-48a1-94ba-0541e5596c5c","Type":"ContainerStarted","Data":"dd5f779a44e0378bbef8e71d300541727499a7027732a574daddc97550b70c99"} Dec 03 13:30:16 crc kubenswrapper[4690]: I1203 13:30:16.587779 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" event={"ID":"4ce1db85-6b25-47fd-84ee-47406aa77eec","Type":"ContainerStarted","Data":"158c41f2268c7b4d37c7ee13ccd0b0cac886b46121bfcd27c7b1519196189141"} Dec 03 13:30:16 crc kubenswrapper[4690]: I1203 13:30:16.824027 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:30:16 crc kubenswrapper[4690]: I1203 13:30:16.824092 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:30:16 crc kubenswrapper[4690]: I1203 13:30:16.824143 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:30:16 crc kubenswrapper[4690]: I1203 13:30:16.824997 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"22570d41f66f8cf71e73ec062157689b8763204a45407561d1852ebc8cede0d1"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:30:16 crc kubenswrapper[4690]: I1203 13:30:16.825056 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://22570d41f66f8cf71e73ec062157689b8763204a45407561d1852ebc8cede0d1" gracePeriod=600 Dec 03 13:30:17 crc kubenswrapper[4690]: I1203 13:30:17.601773 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="22570d41f66f8cf71e73ec062157689b8763204a45407561d1852ebc8cede0d1" exitCode=0 Dec 03 13:30:17 crc kubenswrapper[4690]: I1203 13:30:17.601822 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"22570d41f66f8cf71e73ec062157689b8763204a45407561d1852ebc8cede0d1"} Dec 03 13:30:17 crc kubenswrapper[4690]: I1203 13:30:17.601887 4690 scope.go:117] "RemoveContainer" containerID="7081acf7649c775529e892a032f5a109816cdd926c9cbd0790c89b5ae1594b9f" Dec 03 13:30:23 crc kubenswrapper[4690]: I1203 13:30:23.665345 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"d13f0eb20e5e81edaffaa899e1251f34d5ba5b9191f18b642a8f24f971ecc07d"} Dec 03 13:30:24 crc kubenswrapper[4690]: I1203 13:30:24.755167 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c157eb34-b68f-4510-aafd-0c5e73435efe","Type":"ContainerStarted","Data":"81afa01f4be53c22edaa3557eceea0e78fa47f788908b2fa04f7e7fae69cfbe5"} Dec 03 13:30:24 crc kubenswrapper[4690]: I1203 13:30:24.761232 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e5d47d61-0012-4eb8-ba88-da731d54ad45","Type":"ContainerStarted","Data":"1bce21ee4cfd0655ec74d6e66ea415d39b1d999f23ed54001b794808959596d1"} Dec 03 13:30:24 crc kubenswrapper[4690]: I1203 13:30:24.767617 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c44ebdaa-211b-4ac6-90c6-5236eaee6295","Type":"ContainerStarted","Data":"2e578ce107c6de859200b688081bab9f4a227955864a9293b3f05b0c5d1bb7dc"} Dec 03 13:30:24 crc kubenswrapper[4690]: I1203 13:30:24.769525 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"187a8435-f796-4435-a0b0-211daf924823","Type":"ContainerStarted","Data":"5916118e49b92745d6f37e2df8de28fd4dd4b44d7030491cb910ff753c9994ba"} Dec 03 13:30:24 crc kubenswrapper[4690]: I1203 13:30:24.821259 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=32.433814408 podStartE2EDuration="44.821235836s" podCreationTimestamp="2025-12-03 13:29:40 +0000 UTC" firstStartedPulling="2025-12-03 13:30:07.929503994 +0000 UTC m=+1253.910424427" lastFinishedPulling="2025-12-03 13:30:20.316925412 +0000 UTC m=+1266.297845855" observedRunningTime="2025-12-03 13:30:24.808325835 +0000 UTC m=+1270.789246268" watchObservedRunningTime="2025-12-03 13:30:24.821235836 +0000 UTC m=+1270.802156269" Dec 03 13:30:25 crc kubenswrapper[4690]: I1203 13:30:25.779769 4690 generic.go:334] "Generic (PLEG): container finished" podID="cf553f8a-b61b-469b-9c0c-d602d73ef4c2" containerID="57b50483e2ecdbf53bd5b1134a1f68a37fe0034f76b74c517d4d43065a8aa508" exitCode=0 Dec 03 13:30:25 crc kubenswrapper[4690]: I1203 13:30:25.779845 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-znnxj" event={"ID":"cf553f8a-b61b-469b-9c0c-d602d73ef4c2","Type":"ContainerDied","Data":"57b50483e2ecdbf53bd5b1134a1f68a37fe0034f76b74c517d4d43065a8aa508"} Dec 03 13:30:25 crc kubenswrapper[4690]: I1203 13:30:25.781055 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.791068 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7b78850b-78ce-4668-97e8-d4929857b34a","Type":"ContainerStarted","Data":"c4ad4a3b6b0af4f5c90a1321900f91f4b892c5440d86deb04c585b5f395f4036"} Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.791422 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.793407 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" event={"ID":"09ac08bf-8841-48a1-94ba-0541e5596c5c","Type":"ContainerStarted","Data":"017e90e64b7f8e9d23fa421e21acde00922c858279b6ac2d950209948eba926b"} Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.795197 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" event={"ID":"4ce1db85-6b25-47fd-84ee-47406aa77eec","Type":"ContainerStarted","Data":"123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f"} Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.797236 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c44ebdaa-211b-4ac6-90c6-5236eaee6295","Type":"ContainerStarted","Data":"d427ceb0777fa04ed621ab27e61236778a4e421f2aa66413db13fb695c309b84"} Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.798948 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mf8xw" event={"ID":"64710eed-cdda-4f62-a007-766dd92f0b29","Type":"ContainerStarted","Data":"dfa20d6a54e7ee528caaba8e3372afce3af2a676b5cdd0770e40a38589c04d4e"} Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.799078 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-mf8xw" Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.800799 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-wjlxx" event={"ID":"338580e8-66c9-4169-93f5-1e7f530de31c","Type":"ContainerStarted","Data":"cb1c09f248aa41745a873dff61f75ebce5c05395a855d96a51f025f3fa5e53c5"} Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.803852 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d19eec48-0864-43d8-8071-1aef21a26286","Type":"ContainerStarted","Data":"c1dda51d1fac8aff84b48a5abac5e658d6391c0d5e1b81a715bb2c2ed6972435"} Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.811392 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=26.974498049 podStartE2EDuration="43.811373193s" podCreationTimestamp="2025-12-03 13:29:43 +0000 UTC" firstStartedPulling="2025-12-03 13:30:08.730062459 +0000 UTC m=+1254.710982892" lastFinishedPulling="2025-12-03 13:30:25.566937603 +0000 UTC m=+1271.547858036" observedRunningTime="2025-12-03 13:30:26.806841447 +0000 UTC m=+1272.787761880" watchObservedRunningTime="2025-12-03 13:30:26.811373193 +0000 UTC m=+1272.792293626" Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.842812 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=29.90736447 podStartE2EDuration="41.842794085s" podCreationTimestamp="2025-12-03 13:29:45 +0000 UTC" firstStartedPulling="2025-12-03 13:30:08.708353044 +0000 UTC m=+1254.689273477" lastFinishedPulling="2025-12-03 13:30:20.643782659 +0000 UTC m=+1266.624703092" observedRunningTime="2025-12-03 13:30:26.834155374 +0000 UTC m=+1272.815075807" watchObservedRunningTime="2025-12-03 13:30:26.842794085 +0000 UTC m=+1272.823714518" Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.868977 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-wjlxx" podStartSLOduration=4.579561241 podStartE2EDuration="16.868949775s" podCreationTimestamp="2025-12-03 13:30:10 +0000 UTC" firstStartedPulling="2025-12-03 13:30:11.819182356 +0000 UTC m=+1257.800102789" lastFinishedPulling="2025-12-03 13:30:24.10857089 +0000 UTC m=+1270.089491323" observedRunningTime="2025-12-03 13:30:26.858190394 +0000 UTC m=+1272.839110827" watchObservedRunningTime="2025-12-03 13:30:26.868949775 +0000 UTC m=+1272.849870208" Dec 03 13:30:26 crc kubenswrapper[4690]: I1203 13:30:26.901156 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mf8xw" podStartSLOduration=27.038680885 podStartE2EDuration="40.901136065s" podCreationTimestamp="2025-12-03 13:29:46 +0000 UTC" firstStartedPulling="2025-12-03 13:30:08.846999315 +0000 UTC m=+1254.827919758" lastFinishedPulling="2025-12-03 13:30:22.709454505 +0000 UTC m=+1268.690374938" observedRunningTime="2025-12-03 13:30:26.891761796 +0000 UTC m=+1272.872682239" watchObservedRunningTime="2025-12-03 13:30:26.901136065 +0000 UTC m=+1272.882056498" Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.817165 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-znnxj" event={"ID":"cf553f8a-b61b-469b-9c0c-d602d73ef4c2","Type":"ContainerStarted","Data":"861e4a186ce96bd3176771e86d8357e85ba6032babb998eb55af7c2f27cac54c"} Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.817528 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-znnxj" event={"ID":"cf553f8a-b61b-469b-9c0c-d602d73ef4c2","Type":"ContainerStarted","Data":"1d2445bab54de32109e621dadb8e732dae4d75921f426c35608813ef92fffe39"} Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.817604 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.817640 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.820406 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d19eec48-0864-43d8-8071-1aef21a26286","Type":"ContainerStarted","Data":"7745abd11c83d6c1204e98abaff1ffc39236649c3e49a42ae69d494ae739e336"} Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.823096 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5930069-12cd-4b0b-8aab-0f32ea9057a9","Type":"ContainerStarted","Data":"4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928"} Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.827047 4690 generic.go:334] "Generic (PLEG): container finished" podID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerID="017e90e64b7f8e9d23fa421e21acde00922c858279b6ac2d950209948eba926b" exitCode=0 Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.827476 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" event={"ID":"09ac08bf-8841-48a1-94ba-0541e5596c5c","Type":"ContainerDied","Data":"017e90e64b7f8e9d23fa421e21acde00922c858279b6ac2d950209948eba926b"} Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.836069 4690 generic.go:334] "Generic (PLEG): container finished" podID="4ce1db85-6b25-47fd-84ee-47406aa77eec" containerID="123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f" exitCode=0 Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.836171 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" event={"ID":"4ce1db85-6b25-47fd-84ee-47406aa77eec","Type":"ContainerDied","Data":"123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f"} Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.840264 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6173a433-90f9-4e48-8297-81767d5ce03e","Type":"ContainerStarted","Data":"ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670"} Dec 03 13:30:27 crc kubenswrapper[4690]: I1203 13:30:27.848850 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-znnxj" podStartSLOduration=28.181532056 podStartE2EDuration="41.848827509s" podCreationTimestamp="2025-12-03 13:29:46 +0000 UTC" firstStartedPulling="2025-12-03 13:30:09.042296476 +0000 UTC m=+1255.023216929" lastFinishedPulling="2025-12-03 13:30:22.709591949 +0000 UTC m=+1268.690512382" observedRunningTime="2025-12-03 13:30:27.846167307 +0000 UTC m=+1273.827087750" watchObservedRunningTime="2025-12-03 13:30:27.848827509 +0000 UTC m=+1273.829747952" Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.032550 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=25.295301898 podStartE2EDuration="39.03252638s" podCreationTimestamp="2025-12-03 13:29:49 +0000 UTC" firstStartedPulling="2025-12-03 13:30:08.97337576 +0000 UTC m=+1254.954296213" lastFinishedPulling="2025-12-03 13:30:22.710600262 +0000 UTC m=+1268.691520695" observedRunningTime="2025-12-03 13:30:28.023247354 +0000 UTC m=+1274.004167797" watchObservedRunningTime="2025-12-03 13:30:28.03252638 +0000 UTC m=+1274.013446823" Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.761259 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.802424 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.854577 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" event={"ID":"09ac08bf-8841-48a1-94ba-0541e5596c5c","Type":"ContainerStarted","Data":"cc57b69c4690f514b665dd4236483dc54dcb898e8f00df791fab9a063e1fddae"} Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.854772 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.859605 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" event={"ID":"4ce1db85-6b25-47fd-84ee-47406aa77eec","Type":"ContainerStarted","Data":"7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba"} Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.859658 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.860274 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.884816 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" podStartSLOduration=11.527924861 podStartE2EDuration="18.88478765s" podCreationTimestamp="2025-12-03 13:30:10 +0000 UTC" firstStartedPulling="2025-12-03 13:30:15.959301255 +0000 UTC m=+1261.940221688" lastFinishedPulling="2025-12-03 13:30:23.316164034 +0000 UTC m=+1269.297084477" observedRunningTime="2025-12-03 13:30:28.87837437 +0000 UTC m=+1274.859294813" watchObservedRunningTime="2025-12-03 13:30:28.88478765 +0000 UTC m=+1274.865708093" Dec 03 13:30:28 crc kubenswrapper[4690]: I1203 13:30:28.911320 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" podStartSLOduration=11.168099439 podStartE2EDuration="17.911295517s" podCreationTimestamp="2025-12-03 13:30:11 +0000 UTC" firstStartedPulling="2025-12-03 13:30:15.966077613 +0000 UTC m=+1261.946998046" lastFinishedPulling="2025-12-03 13:30:22.709273691 +0000 UTC m=+1268.690194124" observedRunningTime="2025-12-03 13:30:28.903482695 +0000 UTC m=+1274.884403138" watchObservedRunningTime="2025-12-03 13:30:28.911295517 +0000 UTC m=+1274.892215950" Dec 03 13:30:29 crc kubenswrapper[4690]: I1203 13:30:29.783158 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 13:30:29 crc kubenswrapper[4690]: I1203 13:30:29.830134 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 13:30:29 crc kubenswrapper[4690]: I1203 13:30:29.869496 4690 generic.go:334] "Generic (PLEG): container finished" podID="e5d47d61-0012-4eb8-ba88-da731d54ad45" containerID="1bce21ee4cfd0655ec74d6e66ea415d39b1d999f23ed54001b794808959596d1" exitCode=0 Dec 03 13:30:29 crc kubenswrapper[4690]: I1203 13:30:29.870015 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e5d47d61-0012-4eb8-ba88-da731d54ad45","Type":"ContainerDied","Data":"1bce21ee4cfd0655ec74d6e66ea415d39b1d999f23ed54001b794808959596d1"} Dec 03 13:30:29 crc kubenswrapper[4690]: I1203 13:30:29.874080 4690 generic.go:334] "Generic (PLEG): container finished" podID="c157eb34-b68f-4510-aafd-0c5e73435efe" containerID="81afa01f4be53c22edaa3557eceea0e78fa47f788908b2fa04f7e7fae69cfbe5" exitCode=0 Dec 03 13:30:29 crc kubenswrapper[4690]: I1203 13:30:29.874207 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c157eb34-b68f-4510-aafd-0c5e73435efe","Type":"ContainerDied","Data":"81afa01f4be53c22edaa3557eceea0e78fa47f788908b2fa04f7e7fae69cfbe5"} Dec 03 13:30:29 crc kubenswrapper[4690]: I1203 13:30:29.876103 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 13:30:29 crc kubenswrapper[4690]: I1203 13:30:29.936126 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 13:30:30 crc kubenswrapper[4690]: I1203 13:30:30.883185 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c157eb34-b68f-4510-aafd-0c5e73435efe","Type":"ContainerStarted","Data":"2b80bd2fe52b4560994617795e427a5bc82c738f608e817f598901f6d34dc8fc"} Dec 03 13:30:30 crc kubenswrapper[4690]: I1203 13:30:30.887061 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e5d47d61-0012-4eb8-ba88-da731d54ad45","Type":"ContainerStarted","Data":"10f1a182dd943449daa7699387549e4a85e4f0dee95fa3e5a60e458cac6a1331"} Dec 03 13:30:30 crc kubenswrapper[4690]: I1203 13:30:30.908937 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 13:30:30 crc kubenswrapper[4690]: I1203 13:30:30.923386 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=34.532335007 podStartE2EDuration="51.923363335s" podCreationTimestamp="2025-12-03 13:29:39 +0000 UTC" firstStartedPulling="2025-12-03 13:30:02.926166531 +0000 UTC m=+1248.907086974" lastFinishedPulling="2025-12-03 13:30:20.317194869 +0000 UTC m=+1266.298115302" observedRunningTime="2025-12-03 13:30:30.904749612 +0000 UTC m=+1276.885670045" watchObservedRunningTime="2025-12-03 13:30:30.923363335 +0000 UTC m=+1276.904283768" Dec 03 13:30:30 crc kubenswrapper[4690]: I1203 13:30:30.924282 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=40.096881073 podStartE2EDuration="53.924276247s" podCreationTimestamp="2025-12-03 13:29:37 +0000 UTC" firstStartedPulling="2025-12-03 13:30:08.570229075 +0000 UTC m=+1254.551149508" lastFinishedPulling="2025-12-03 13:30:22.397624249 +0000 UTC m=+1268.378544682" observedRunningTime="2025-12-03 13:30:30.920753025 +0000 UTC m=+1276.901673468" watchObservedRunningTime="2025-12-03 13:30:30.924276247 +0000 UTC m=+1276.905196680" Dec 03 13:30:31 crc kubenswrapper[4690]: I1203 13:30:31.934830 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.102263 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 13:30:32 crc kubenswrapper[4690]: E1203 13:30:32.103015 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9edbc53-d827-45be-aa22-bb691422fb4d" containerName="collect-profiles" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.103033 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9edbc53-d827-45be-aa22-bb691422fb4d" containerName="collect-profiles" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.103237 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9edbc53-d827-45be-aa22-bb691422fb4d" containerName="collect-profiles" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.104257 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.106734 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.107153 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-ztg5d" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.107408 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.108100 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.118188 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.196160 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mljp\" (UniqueName: \"kubernetes.io/projected/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-kube-api-access-7mljp\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.196293 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-scripts\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.196328 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.196400 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-config\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.196432 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.196521 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.196544 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.298679 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-config\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.298747 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.298799 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.298820 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.298926 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mljp\" (UniqueName: \"kubernetes.io/projected/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-kube-api-access-7mljp\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.298999 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-scripts\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.299028 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.300225 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.300491 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-scripts\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.300585 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-config\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.307445 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.323844 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.325613 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.326954 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mljp\" (UniqueName: \"kubernetes.io/projected/a172f1ad-a0fb-4f37-92e0-2c7267d11fa8-kube-api-access-7mljp\") pod \"ovn-northd-0\" (UID: \"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8\") " pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.434213 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.877442 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 13:30:32 crc kubenswrapper[4690]: W1203 13:30:32.884682 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda172f1ad_a0fb_4f37_92e0_2c7267d11fa8.slice/crio-feb214ffcb31bca9ca6b80a0c03474ba19abdc0739b8230a41470b1fd7ddfb9d WatchSource:0}: Error finding container feb214ffcb31bca9ca6b80a0c03474ba19abdc0739b8230a41470b1fd7ddfb9d: Status 404 returned error can't find the container with id feb214ffcb31bca9ca6b80a0c03474ba19abdc0739b8230a41470b1fd7ddfb9d Dec 03 13:30:32 crc kubenswrapper[4690]: I1203 13:30:32.907706 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8","Type":"ContainerStarted","Data":"feb214ffcb31bca9ca6b80a0c03474ba19abdc0739b8230a41470b1fd7ddfb9d"} Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.388252 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.538130 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-h2hnf"] Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.538544 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" podUID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerName="dnsmasq-dns" containerID="cri-o://cc57b69c4690f514b665dd4236483dc54dcb898e8f00df791fab9a063e1fddae" gracePeriod=10 Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.542099 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.597051 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-qtkhc"] Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.598953 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.634834 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qtkhc"] Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.739119 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-dns-svc\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.739179 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.739229 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgs72\" (UniqueName: \"kubernetes.io/projected/aed37082-28cb-49e4-9d87-e8246df4b55c-kube-api-access-jgs72\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.739439 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.739616 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-config\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.841359 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-dns-svc\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.841961 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.842095 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgs72\" (UniqueName: \"kubernetes.io/projected/aed37082-28cb-49e4-9d87-e8246df4b55c-kube-api-access-jgs72\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.842254 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.842365 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-config\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.842633 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-dns-svc\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.843229 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-config\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.843465 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.843935 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.895289 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgs72\" (UniqueName: \"kubernetes.io/projected/aed37082-28cb-49e4-9d87-e8246df4b55c-kube-api-access-jgs72\") pod \"dnsmasq-dns-698758b865-qtkhc\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:33 crc kubenswrapper[4690]: I1203 13:30:33.947411 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.456685 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qtkhc"] Dec 03 13:30:34 crc kubenswrapper[4690]: W1203 13:30:34.458234 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaed37082_28cb_49e4_9d87_e8246df4b55c.slice/crio-dc1bf79a1e689f1d33e78d95502bd9485e7a3b6081cfa3986a0465ea1f232c9d WatchSource:0}: Error finding container dc1bf79a1e689f1d33e78d95502bd9485e7a3b6081cfa3986a0465ea1f232c9d: Status 404 returned error can't find the container with id dc1bf79a1e689f1d33e78d95502bd9485e7a3b6081cfa3986a0465ea1f232c9d Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.567284 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.573243 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.575623 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.575988 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.576177 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.576348 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-n4j2m" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.590785 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.656923 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/28748086-659b-453c-bb91-8d9204374da6-cache\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.657100 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.657205 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pdgk\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-kube-api-access-8pdgk\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.657236 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/28748086-659b-453c-bb91-8d9204374da6-lock\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.657305 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.763249 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pdgk\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-kube-api-access-8pdgk\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.763316 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/28748086-659b-453c-bb91-8d9204374da6-lock\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.763377 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.763433 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/28748086-659b-453c-bb91-8d9204374da6-cache\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.763513 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.763941 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.764936 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/28748086-659b-453c-bb91-8d9204374da6-lock\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: E1203 13:30:34.765320 4690 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 13:30:34 crc kubenswrapper[4690]: E1203 13:30:34.765341 4690 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 13:30:34 crc kubenswrapper[4690]: E1203 13:30:34.765380 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift podName:28748086-659b-453c-bb91-8d9204374da6 nodeName:}" failed. No retries permitted until 2025-12-03 13:30:35.265364066 +0000 UTC m=+1281.246284499 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift") pod "swift-storage-0" (UID: "28748086-659b-453c-bb91-8d9204374da6") : configmap "swift-ring-files" not found Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.765704 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/28748086-659b-453c-bb91-8d9204374da6-cache\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.818808 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pdgk\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-kube-api-access-8pdgk\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.819057 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:34 crc kubenswrapper[4690]: I1203 13:30:34.982022 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qtkhc" event={"ID":"aed37082-28cb-49e4-9d87-e8246df4b55c","Type":"ContainerStarted","Data":"dc1bf79a1e689f1d33e78d95502bd9485e7a3b6081cfa3986a0465ea1f232c9d"} Dec 03 13:30:35 crc kubenswrapper[4690]: I1203 13:30:35.272466 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:35 crc kubenswrapper[4690]: E1203 13:30:35.272728 4690 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 13:30:35 crc kubenswrapper[4690]: E1203 13:30:35.272955 4690 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 13:30:35 crc kubenswrapper[4690]: E1203 13:30:35.273027 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift podName:28748086-659b-453c-bb91-8d9204374da6 nodeName:}" failed. No retries permitted until 2025-12-03 13:30:36.273007276 +0000 UTC m=+1282.253927709 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift") pod "swift-storage-0" (UID: "28748086-659b-453c-bb91-8d9204374da6") : configmap "swift-ring-files" not found Dec 03 13:30:35 crc kubenswrapper[4690]: I1203 13:30:35.994505 4690 generic.go:334] "Generic (PLEG): container finished" podID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerID="cc57b69c4690f514b665dd4236483dc54dcb898e8f00df791fab9a063e1fddae" exitCode=0 Dec 03 13:30:35 crc kubenswrapper[4690]: I1203 13:30:35.994605 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" event={"ID":"09ac08bf-8841-48a1-94ba-0541e5596c5c","Type":"ContainerDied","Data":"cc57b69c4690f514b665dd4236483dc54dcb898e8f00df791fab9a063e1fddae"} Dec 03 13:30:36 crc kubenswrapper[4690]: I1203 13:30:36.216579 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" podUID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 03 13:30:36 crc kubenswrapper[4690]: I1203 13:30:36.294207 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:36 crc kubenswrapper[4690]: E1203 13:30:36.294404 4690 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 13:30:36 crc kubenswrapper[4690]: E1203 13:30:36.294425 4690 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 13:30:36 crc kubenswrapper[4690]: E1203 13:30:36.294484 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift podName:28748086-659b-453c-bb91-8d9204374da6 nodeName:}" failed. No retries permitted until 2025-12-03 13:30:38.294466988 +0000 UTC m=+1284.275387421 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift") pod "swift-storage-0" (UID: "28748086-659b-453c-bb91-8d9204374da6") : configmap "swift-ring-files" not found Dec 03 13:30:36 crc kubenswrapper[4690]: I1203 13:30:36.587336 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.336397 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:38 crc kubenswrapper[4690]: E1203 13:30:38.336774 4690 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 13:30:38 crc kubenswrapper[4690]: E1203 13:30:38.336898 4690 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 13:30:38 crc kubenswrapper[4690]: E1203 13:30:38.336975 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift podName:28748086-659b-453c-bb91-8d9204374da6 nodeName:}" failed. No retries permitted until 2025-12-03 13:30:42.336954925 +0000 UTC m=+1288.317875358 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift") pod "swift-storage-0" (UID: "28748086-659b-453c-bb91-8d9204374da6") : configmap "swift-ring-files" not found Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.565172 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bl55c"] Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.569707 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.572815 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.572925 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.573024 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.591071 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bl55c"] Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.644310 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-dispersionconf\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.644360 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-combined-ca-bundle\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.644401 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-scripts\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.644449 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-ring-data-devices\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.644465 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/262bf571-0151-4c11-a05c-1270656144f9-etc-swift\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.644495 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwt2b\" (UniqueName: \"kubernetes.io/projected/262bf571-0151-4c11-a05c-1270656144f9-kube-api-access-xwt2b\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.644588 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-swiftconf\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.721373 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.746477 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-dispersionconf\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.746560 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-combined-ca-bundle\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.746614 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-scripts\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.746666 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-ring-data-devices\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.746690 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/262bf571-0151-4c11-a05c-1270656144f9-etc-swift\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.746737 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwt2b\" (UniqueName: \"kubernetes.io/projected/262bf571-0151-4c11-a05c-1270656144f9-kube-api-access-xwt2b\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.746808 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-swiftconf\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.748432 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-ring-data-devices\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.749124 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/262bf571-0151-4c11-a05c-1270656144f9-etc-swift\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.749543 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-scripts\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.753459 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-dispersionconf\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.753609 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-combined-ca-bundle\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.761333 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-swiftconf\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.763622 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwt2b\" (UniqueName: \"kubernetes.io/projected/262bf571-0151-4c11-a05c-1270656144f9-kube-api-access-xwt2b\") pod \"swift-ring-rebalance-bl55c\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.849375 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-config\") pod \"09ac08bf-8841-48a1-94ba-0541e5596c5c\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.849487 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kpxw\" (UniqueName: \"kubernetes.io/projected/09ac08bf-8841-48a1-94ba-0541e5596c5c-kube-api-access-2kpxw\") pod \"09ac08bf-8841-48a1-94ba-0541e5596c5c\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.849551 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-dns-svc\") pod \"09ac08bf-8841-48a1-94ba-0541e5596c5c\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.849599 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-ovsdbserver-nb\") pod \"09ac08bf-8841-48a1-94ba-0541e5596c5c\" (UID: \"09ac08bf-8841-48a1-94ba-0541e5596c5c\") " Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.854560 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ac08bf-8841-48a1-94ba-0541e5596c5c-kube-api-access-2kpxw" (OuterVolumeSpecName: "kube-api-access-2kpxw") pod "09ac08bf-8841-48a1-94ba-0541e5596c5c" (UID: "09ac08bf-8841-48a1-94ba-0541e5596c5c"). InnerVolumeSpecName "kube-api-access-2kpxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.892556 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.896945 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-config" (OuterVolumeSpecName: "config") pod "09ac08bf-8841-48a1-94ba-0541e5596c5c" (UID: "09ac08bf-8841-48a1-94ba-0541e5596c5c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.908587 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "09ac08bf-8841-48a1-94ba-0541e5596c5c" (UID: "09ac08bf-8841-48a1-94ba-0541e5596c5c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.912263 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "09ac08bf-8841-48a1-94ba-0541e5596c5c" (UID: "09ac08bf-8841-48a1-94ba-0541e5596c5c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.952059 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.952104 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kpxw\" (UniqueName: \"kubernetes.io/projected/09ac08bf-8841-48a1-94ba-0541e5596c5c-kube-api-access-2kpxw\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.952121 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:38 crc kubenswrapper[4690]: I1203 13:30:38.952166 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09ac08bf-8841-48a1-94ba-0541e5596c5c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.027002 4690 generic.go:334] "Generic (PLEG): container finished" podID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerID="cc10549ac276b05176be491d8718bf387f431c1346f8f41c57001a377a09f8a0" exitCode=0 Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.027062 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qtkhc" event={"ID":"aed37082-28cb-49e4-9d87-e8246df4b55c","Type":"ContainerDied","Data":"cc10549ac276b05176be491d8718bf387f431c1346f8f41c57001a377a09f8a0"} Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.030659 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" event={"ID":"09ac08bf-8841-48a1-94ba-0541e5596c5c","Type":"ContainerDied","Data":"dd5f779a44e0378bbef8e71d300541727499a7027732a574daddc97550b70c99"} Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.030692 4690 scope.go:117] "RemoveContainer" containerID="cc57b69c4690f514b665dd4236483dc54dcb898e8f00df791fab9a063e1fddae" Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.030732 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-h2hnf" Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.075405 4690 scope.go:117] "RemoveContainer" containerID="017e90e64b7f8e9d23fa421e21acde00922c858279b6ac2d950209948eba926b" Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.086609 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-h2hnf"] Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.097183 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-h2hnf"] Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.180903 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.180955 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 13:30:39 crc kubenswrapper[4690]: I1203 13:30:39.346766 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bl55c"] Dec 03 13:30:39 crc kubenswrapper[4690]: W1203 13:30:39.715987 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod262bf571_0151_4c11_a05c_1270656144f9.slice/crio-19eea06d789dd49925ccf86d413227fcec4a4602763d473be48e370089f1a8ba WatchSource:0}: Error finding container 19eea06d789dd49925ccf86d413227fcec4a4602763d473be48e370089f1a8ba: Status 404 returned error can't find the container with id 19eea06d789dd49925ccf86d413227fcec4a4602763d473be48e370089f1a8ba Dec 03 13:30:40 crc kubenswrapper[4690]: I1203 13:30:40.044616 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bl55c" event={"ID":"262bf571-0151-4c11-a05c-1270656144f9","Type":"ContainerStarted","Data":"19eea06d789dd49925ccf86d413227fcec4a4602763d473be48e370089f1a8ba"} Dec 03 13:30:40 crc kubenswrapper[4690]: I1203 13:30:40.330681 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ac08bf-8841-48a1-94ba-0541e5596c5c" path="/var/lib/kubelet/pods/09ac08bf-8841-48a1-94ba-0541e5596c5c/volumes" Dec 03 13:30:40 crc kubenswrapper[4690]: I1203 13:30:40.556003 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 13:30:40 crc kubenswrapper[4690]: I1203 13:30:40.625627 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="e5d47d61-0012-4eb8-ba88-da731d54ad45" containerName="galera" probeResult="failure" output=< Dec 03 13:30:40 crc kubenswrapper[4690]: wsrep_local_state_comment (Joined) differs from Synced Dec 03 13:30:40 crc kubenswrapper[4690]: > Dec 03 13:30:40 crc kubenswrapper[4690]: I1203 13:30:40.725503 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 13:30:40 crc kubenswrapper[4690]: I1203 13:30:40.726027 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 13:30:41 crc kubenswrapper[4690]: I1203 13:30:41.066057 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qtkhc" event={"ID":"aed37082-28cb-49e4-9d87-e8246df4b55c","Type":"ContainerStarted","Data":"54660f3c082858a9860457fc52b1df5e7817a7957a36fb967cb370ed62948347"} Dec 03 13:30:41 crc kubenswrapper[4690]: I1203 13:30:41.067958 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:41 crc kubenswrapper[4690]: I1203 13:30:41.072179 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8","Type":"ContainerStarted","Data":"e82e0988e1f81b2e2e2f34c3a2e3eb9f91693519843c8d7a90db35e2529af749"} Dec 03 13:30:41 crc kubenswrapper[4690]: I1203 13:30:41.091903 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-qtkhc" podStartSLOduration=8.091880803 podStartE2EDuration="8.091880803s" podCreationTimestamp="2025-12-03 13:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:30:41.084687746 +0000 UTC m=+1287.065608189" watchObservedRunningTime="2025-12-03 13:30:41.091880803 +0000 UTC m=+1287.072801246" Dec 03 13:30:42 crc kubenswrapper[4690]: I1203 13:30:42.083667 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a172f1ad-a0fb-4f37-92e0-2c7267d11fa8","Type":"ContainerStarted","Data":"42e64209b1624bf999deb9b327229426528f52150ba93fce5590515ec07e4dad"} Dec 03 13:30:42 crc kubenswrapper[4690]: I1203 13:30:42.084043 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 13:30:42 crc kubenswrapper[4690]: I1203 13:30:42.106082 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.293330036 podStartE2EDuration="10.106059987s" podCreationTimestamp="2025-12-03 13:30:32 +0000 UTC" firstStartedPulling="2025-12-03 13:30:32.887924735 +0000 UTC m=+1278.868845168" lastFinishedPulling="2025-12-03 13:30:40.700654686 +0000 UTC m=+1286.681575119" observedRunningTime="2025-12-03 13:30:42.100144329 +0000 UTC m=+1288.081064772" watchObservedRunningTime="2025-12-03 13:30:42.106059987 +0000 UTC m=+1288.086980420" Dec 03 13:30:42 crc kubenswrapper[4690]: I1203 13:30:42.435261 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:42 crc kubenswrapper[4690]: E1203 13:30:42.435542 4690 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 13:30:42 crc kubenswrapper[4690]: E1203 13:30:42.435556 4690 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 13:30:42 crc kubenswrapper[4690]: E1203 13:30:42.435604 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift podName:28748086-659b-453c-bb91-8d9204374da6 nodeName:}" failed. No retries permitted until 2025-12-03 13:30:50.435588946 +0000 UTC m=+1296.416509379 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift") pod "swift-storage-0" (UID: "28748086-659b-453c-bb91-8d9204374da6") : configmap "swift-ring-files" not found Dec 03 13:30:43 crc kubenswrapper[4690]: I1203 13:30:43.954460 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 13:30:44 crc kubenswrapper[4690]: I1203 13:30:44.048787 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 13:30:46 crc kubenswrapper[4690]: I1203 13:30:46.118761 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bl55c" event={"ID":"262bf571-0151-4c11-a05c-1270656144f9","Type":"ContainerStarted","Data":"9a04bd0b28e1da640266c690eeca3c69b4ef7a54cb18526c3d6aae705a4eec47"} Dec 03 13:30:46 crc kubenswrapper[4690]: I1203 13:30:46.144369 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bl55c" podStartSLOduration=2.470286448 podStartE2EDuration="8.144347032s" podCreationTimestamp="2025-12-03 13:30:38 +0000 UTC" firstStartedPulling="2025-12-03 13:30:39.723745131 +0000 UTC m=+1285.704665554" lastFinishedPulling="2025-12-03 13:30:45.397805705 +0000 UTC m=+1291.378726138" observedRunningTime="2025-12-03 13:30:46.139648832 +0000 UTC m=+1292.120569275" watchObservedRunningTime="2025-12-03 13:30:46.144347032 +0000 UTC m=+1292.125267465" Dec 03 13:30:48 crc kubenswrapper[4690]: I1203 13:30:48.950077 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:30:49 crc kubenswrapper[4690]: I1203 13:30:49.012808 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kvkvr"] Dec 03 13:30:49 crc kubenswrapper[4690]: I1203 13:30:49.013409 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" podUID="4ce1db85-6b25-47fd-84ee-47406aa77eec" containerName="dnsmasq-dns" containerID="cri-o://7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba" gracePeriod=10 Dec 03 13:30:49 crc kubenswrapper[4690]: I1203 13:30:49.258259 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.127350 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.167326 4690 generic.go:334] "Generic (PLEG): container finished" podID="4ce1db85-6b25-47fd-84ee-47406aa77eec" containerID="7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba" exitCode=0 Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.167434 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.167532 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" event={"ID":"4ce1db85-6b25-47fd-84ee-47406aa77eec","Type":"ContainerDied","Data":"7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba"} Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.168741 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kvkvr" event={"ID":"4ce1db85-6b25-47fd-84ee-47406aa77eec","Type":"ContainerDied","Data":"158c41f2268c7b4d37c7ee13ccd0b0cac886b46121bfcd27c7b1519196189141"} Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.168788 4690 scope.go:117] "RemoveContainer" containerID="7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.192769 4690 scope.go:117] "RemoveContainer" containerID="123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.213887 4690 scope.go:117] "RemoveContainer" containerID="7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba" Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.214393 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba\": container with ID starting with 7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba not found: ID does not exist" containerID="7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.214443 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba"} err="failed to get container status \"7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba\": rpc error: code = NotFound desc = could not find container \"7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba\": container with ID starting with 7f3e0845a5f54fde53c07379990aa3bcb423e3692614dc2effa55eb28ff06cba not found: ID does not exist" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.214475 4690 scope.go:117] "RemoveContainer" containerID="123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f" Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.215010 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f\": container with ID starting with 123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f not found: ID does not exist" containerID="123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.215156 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f"} err="failed to get container status \"123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f\": rpc error: code = NotFound desc = could not find container \"123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f\": container with ID starting with 123c2ad0cdcae10a2e18d3506aeb7bbec1121014c85e16aeb250f75903e8966f not found: ID does not exist" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.303568 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-config\") pod \"4ce1db85-6b25-47fd-84ee-47406aa77eec\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.304021 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-dns-svc\") pod \"4ce1db85-6b25-47fd-84ee-47406aa77eec\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.304192 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-sb\") pod \"4ce1db85-6b25-47fd-84ee-47406aa77eec\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.304249 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl9x9\" (UniqueName: \"kubernetes.io/projected/4ce1db85-6b25-47fd-84ee-47406aa77eec-kube-api-access-hl9x9\") pod \"4ce1db85-6b25-47fd-84ee-47406aa77eec\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.304281 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-nb\") pod \"4ce1db85-6b25-47fd-84ee-47406aa77eec\" (UID: \"4ce1db85-6b25-47fd-84ee-47406aa77eec\") " Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.310194 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ce1db85-6b25-47fd-84ee-47406aa77eec-kube-api-access-hl9x9" (OuterVolumeSpecName: "kube-api-access-hl9x9") pod "4ce1db85-6b25-47fd-84ee-47406aa77eec" (UID: "4ce1db85-6b25-47fd-84ee-47406aa77eec"). InnerVolumeSpecName "kube-api-access-hl9x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.347410 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ce1db85-6b25-47fd-84ee-47406aa77eec" (UID: "4ce1db85-6b25-47fd-84ee-47406aa77eec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.375312 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4ce1db85-6b25-47fd-84ee-47406aa77eec" (UID: "4ce1db85-6b25-47fd-84ee-47406aa77eec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.376937 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ce1db85-6b25-47fd-84ee-47406aa77eec" (UID: "4ce1db85-6b25-47fd-84ee-47406aa77eec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.377772 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-config" (OuterVolumeSpecName: "config") pod "4ce1db85-6b25-47fd-84ee-47406aa77eec" (UID: "4ce1db85-6b25-47fd-84ee-47406aa77eec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.407323 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.407366 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl9x9\" (UniqueName: \"kubernetes.io/projected/4ce1db85-6b25-47fd-84ee-47406aa77eec-kube-api-access-hl9x9\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.407386 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.407402 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.407417 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ce1db85-6b25-47fd-84ee-47406aa77eec-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.505768 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kvkvr"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.509322 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.509593 4690 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.509610 4690 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.509659 4690 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift podName:28748086-659b-453c-bb91-8d9204374da6 nodeName:}" failed. No retries permitted until 2025-12-03 13:31:06.509643816 +0000 UTC m=+1312.490564249 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift") pod "swift-storage-0" (UID: "28748086-659b-453c-bb91-8d9204374da6") : configmap "swift-ring-files" not found Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.514768 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kvkvr"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.595345 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8784-account-create-update-nc4fz"] Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.595981 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerName="dnsmasq-dns" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.596014 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerName="dnsmasq-dns" Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.596047 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce1db85-6b25-47fd-84ee-47406aa77eec" containerName="dnsmasq-dns" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.596055 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce1db85-6b25-47fd-84ee-47406aa77eec" containerName="dnsmasq-dns" Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.596083 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerName="init" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.596092 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerName="init" Dec 03 13:30:50 crc kubenswrapper[4690]: E1203 13:30:50.596104 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce1db85-6b25-47fd-84ee-47406aa77eec" containerName="init" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.596111 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce1db85-6b25-47fd-84ee-47406aa77eec" containerName="init" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.596334 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce1db85-6b25-47fd-84ee-47406aa77eec" containerName="dnsmasq-dns" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.596363 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ac08bf-8841-48a1-94ba-0541e5596c5c" containerName="dnsmasq-dns" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.597327 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.606221 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-wn8pw"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.606308 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.607326 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.618355 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8784-account-create-update-nc4fz"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.618409 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wn8pw"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.738702 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssl2g\" (UniqueName: \"kubernetes.io/projected/6bb598ea-0a2a-49b6-a989-533ab58cd524-kube-api-access-ssl2g\") pod \"keystone-db-create-wn8pw\" (UID: \"6bb598ea-0a2a-49b6-a989-533ab58cd524\") " pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.738765 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xnsf\" (UniqueName: \"kubernetes.io/projected/9cf94540-efe2-4e03-80ff-05dbef2b7566-kube-api-access-7xnsf\") pod \"keystone-8784-account-create-update-nc4fz\" (UID: \"9cf94540-efe2-4e03-80ff-05dbef2b7566\") " pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.738800 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf94540-efe2-4e03-80ff-05dbef2b7566-operator-scripts\") pod \"keystone-8784-account-create-update-nc4fz\" (UID: \"9cf94540-efe2-4e03-80ff-05dbef2b7566\") " pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.739130 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb598ea-0a2a-49b6-a989-533ab58cd524-operator-scripts\") pod \"keystone-db-create-wn8pw\" (UID: \"6bb598ea-0a2a-49b6-a989-533ab58cd524\") " pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.792097 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5bb56"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.794121 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5bb56" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.801812 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5bb56"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.840822 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssl2g\" (UniqueName: \"kubernetes.io/projected/6bb598ea-0a2a-49b6-a989-533ab58cd524-kube-api-access-ssl2g\") pod \"keystone-db-create-wn8pw\" (UID: \"6bb598ea-0a2a-49b6-a989-533ab58cd524\") " pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.840909 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xnsf\" (UniqueName: \"kubernetes.io/projected/9cf94540-efe2-4e03-80ff-05dbef2b7566-kube-api-access-7xnsf\") pod \"keystone-8784-account-create-update-nc4fz\" (UID: \"9cf94540-efe2-4e03-80ff-05dbef2b7566\") " pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.840975 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf94540-efe2-4e03-80ff-05dbef2b7566-operator-scripts\") pod \"keystone-8784-account-create-update-nc4fz\" (UID: \"9cf94540-efe2-4e03-80ff-05dbef2b7566\") " pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.841354 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb598ea-0a2a-49b6-a989-533ab58cd524-operator-scripts\") pod \"keystone-db-create-wn8pw\" (UID: \"6bb598ea-0a2a-49b6-a989-533ab58cd524\") " pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.842119 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf94540-efe2-4e03-80ff-05dbef2b7566-operator-scripts\") pod \"keystone-8784-account-create-update-nc4fz\" (UID: \"9cf94540-efe2-4e03-80ff-05dbef2b7566\") " pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.842501 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb598ea-0a2a-49b6-a989-533ab58cd524-operator-scripts\") pod \"keystone-db-create-wn8pw\" (UID: \"6bb598ea-0a2a-49b6-a989-533ab58cd524\") " pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.859806 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xnsf\" (UniqueName: \"kubernetes.io/projected/9cf94540-efe2-4e03-80ff-05dbef2b7566-kube-api-access-7xnsf\") pod \"keystone-8784-account-create-update-nc4fz\" (UID: \"9cf94540-efe2-4e03-80ff-05dbef2b7566\") " pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.860171 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssl2g\" (UniqueName: \"kubernetes.io/projected/6bb598ea-0a2a-49b6-a989-533ab58cd524-kube-api-access-ssl2g\") pod \"keystone-db-create-wn8pw\" (UID: \"6bb598ea-0a2a-49b6-a989-533ab58cd524\") " pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.900246 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a217-account-create-update-9pg4q"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.902410 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.909363 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a217-account-create-update-9pg4q"] Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.909452 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.943901 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cftf9\" (UniqueName: \"kubernetes.io/projected/3ca950cf-73a9-47ae-b07c-b8806057e24b-kube-api-access-cftf9\") pod \"placement-db-create-5bb56\" (UID: \"3ca950cf-73a9-47ae-b07c-b8806057e24b\") " pod="openstack/placement-db-create-5bb56" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.943972 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ca950cf-73a9-47ae-b07c-b8806057e24b-operator-scripts\") pod \"placement-db-create-5bb56\" (UID: \"3ca950cf-73a9-47ae-b07c-b8806057e24b\") " pod="openstack/placement-db-create-5bb56" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.961314 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:50 crc kubenswrapper[4690]: I1203 13:30:50.970576 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.045519 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ca950cf-73a9-47ae-b07c-b8806057e24b-operator-scripts\") pod \"placement-db-create-5bb56\" (UID: \"3ca950cf-73a9-47ae-b07c-b8806057e24b\") " pod="openstack/placement-db-create-5bb56" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.045675 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac7e979-c496-4a3a-8cf0-d3a34d659b85-operator-scripts\") pod \"placement-a217-account-create-update-9pg4q\" (UID: \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\") " pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.045733 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rrx6\" (UniqueName: \"kubernetes.io/projected/aac7e979-c496-4a3a-8cf0-d3a34d659b85-kube-api-access-6rrx6\") pod \"placement-a217-account-create-update-9pg4q\" (UID: \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\") " pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.045770 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cftf9\" (UniqueName: \"kubernetes.io/projected/3ca950cf-73a9-47ae-b07c-b8806057e24b-kube-api-access-cftf9\") pod \"placement-db-create-5bb56\" (UID: \"3ca950cf-73a9-47ae-b07c-b8806057e24b\") " pod="openstack/placement-db-create-5bb56" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.046751 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ca950cf-73a9-47ae-b07c-b8806057e24b-operator-scripts\") pod \"placement-db-create-5bb56\" (UID: \"3ca950cf-73a9-47ae-b07c-b8806057e24b\") " pod="openstack/placement-db-create-5bb56" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.065045 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-rf676"] Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.066623 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rf676" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.087885 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cftf9\" (UniqueName: \"kubernetes.io/projected/3ca950cf-73a9-47ae-b07c-b8806057e24b-kube-api-access-cftf9\") pod \"placement-db-create-5bb56\" (UID: \"3ca950cf-73a9-47ae-b07c-b8806057e24b\") " pod="openstack/placement-db-create-5bb56" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.097898 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rf676"] Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.114960 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5bb56" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.148694 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac7e979-c496-4a3a-8cf0-d3a34d659b85-operator-scripts\") pod \"placement-a217-account-create-update-9pg4q\" (UID: \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\") " pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.148771 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rrx6\" (UniqueName: \"kubernetes.io/projected/aac7e979-c496-4a3a-8cf0-d3a34d659b85-kube-api-access-6rrx6\") pod \"placement-a217-account-create-update-9pg4q\" (UID: \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\") " pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.150243 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac7e979-c496-4a3a-8cf0-d3a34d659b85-operator-scripts\") pod \"placement-a217-account-create-update-9pg4q\" (UID: \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\") " pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.174759 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rrx6\" (UniqueName: \"kubernetes.io/projected/aac7e979-c496-4a3a-8cf0-d3a34d659b85-kube-api-access-6rrx6\") pod \"placement-a217-account-create-update-9pg4q\" (UID: \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\") " pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.235385 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.247107 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-deae-account-create-update-49dp2"] Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.248491 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.250389 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz8vc\" (UniqueName: \"kubernetes.io/projected/9c88c34d-c55a-4456-9f8b-daf83b54896a-kube-api-access-dz8vc\") pod \"glance-db-create-rf676\" (UID: \"9c88c34d-c55a-4456-9f8b-daf83b54896a\") " pod="openstack/glance-db-create-rf676" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.250442 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c88c34d-c55a-4456-9f8b-daf83b54896a-operator-scripts\") pod \"glance-db-create-rf676\" (UID: \"9c88c34d-c55a-4456-9f8b-daf83b54896a\") " pod="openstack/glance-db-create-rf676" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.253614 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.273959 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-deae-account-create-update-49dp2"] Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.352207 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kdsm\" (UniqueName: \"kubernetes.io/projected/54bfb9d7-bab2-4541-abd5-96f850bffdde-kube-api-access-5kdsm\") pod \"glance-deae-account-create-update-49dp2\" (UID: \"54bfb9d7-bab2-4541-abd5-96f850bffdde\") " pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.352579 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz8vc\" (UniqueName: \"kubernetes.io/projected/9c88c34d-c55a-4456-9f8b-daf83b54896a-kube-api-access-dz8vc\") pod \"glance-db-create-rf676\" (UID: \"9c88c34d-c55a-4456-9f8b-daf83b54896a\") " pod="openstack/glance-db-create-rf676" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.352636 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c88c34d-c55a-4456-9f8b-daf83b54896a-operator-scripts\") pod \"glance-db-create-rf676\" (UID: \"9c88c34d-c55a-4456-9f8b-daf83b54896a\") " pod="openstack/glance-db-create-rf676" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.352655 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bfb9d7-bab2-4541-abd5-96f850bffdde-operator-scripts\") pod \"glance-deae-account-create-update-49dp2\" (UID: \"54bfb9d7-bab2-4541-abd5-96f850bffdde\") " pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.353756 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c88c34d-c55a-4456-9f8b-daf83b54896a-operator-scripts\") pod \"glance-db-create-rf676\" (UID: \"9c88c34d-c55a-4456-9f8b-daf83b54896a\") " pod="openstack/glance-db-create-rf676" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.380280 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz8vc\" (UniqueName: \"kubernetes.io/projected/9c88c34d-c55a-4456-9f8b-daf83b54896a-kube-api-access-dz8vc\") pod \"glance-db-create-rf676\" (UID: \"9c88c34d-c55a-4456-9f8b-daf83b54896a\") " pod="openstack/glance-db-create-rf676" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.391460 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rf676" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.454206 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kdsm\" (UniqueName: \"kubernetes.io/projected/54bfb9d7-bab2-4541-abd5-96f850bffdde-kube-api-access-5kdsm\") pod \"glance-deae-account-create-update-49dp2\" (UID: \"54bfb9d7-bab2-4541-abd5-96f850bffdde\") " pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.454314 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bfb9d7-bab2-4541-abd5-96f850bffdde-operator-scripts\") pod \"glance-deae-account-create-update-49dp2\" (UID: \"54bfb9d7-bab2-4541-abd5-96f850bffdde\") " pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.456940 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bfb9d7-bab2-4541-abd5-96f850bffdde-operator-scripts\") pod \"glance-deae-account-create-update-49dp2\" (UID: \"54bfb9d7-bab2-4541-abd5-96f850bffdde\") " pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.486289 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kdsm\" (UniqueName: \"kubernetes.io/projected/54bfb9d7-bab2-4541-abd5-96f850bffdde-kube-api-access-5kdsm\") pod \"glance-deae-account-create-update-49dp2\" (UID: \"54bfb9d7-bab2-4541-abd5-96f850bffdde\") " pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.553478 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8784-account-create-update-nc4fz"] Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.652361 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wn8pw"] Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.664489 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.774679 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5bb56"] Dec 03 13:30:51 crc kubenswrapper[4690]: I1203 13:30:51.826788 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a217-account-create-update-9pg4q"] Dec 03 13:30:52 crc kubenswrapper[4690]: W1203 13:30:52.242334 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cf94540_efe2_4e03_80ff_05dbef2b7566.slice/crio-a88e5e017aa0749291406f813e4389187c846a77e13808e6d3d8ab5a2abccfde WatchSource:0}: Error finding container a88e5e017aa0749291406f813e4389187c846a77e13808e6d3d8ab5a2abccfde: Status 404 returned error can't find the container with id a88e5e017aa0749291406f813e4389187c846a77e13808e6d3d8ab5a2abccfde Dec 03 13:30:52 crc kubenswrapper[4690]: W1203 13:30:52.247055 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bb598ea_0a2a_49b6_a989_533ab58cd524.slice/crio-885e59b01771d1e4ee3f8e7b108370d1c442749c2a436406d952107d7920fd0d WatchSource:0}: Error finding container 885e59b01771d1e4ee3f8e7b108370d1c442749c2a436406d952107d7920fd0d: Status 404 returned error can't find the container with id 885e59b01771d1e4ee3f8e7b108370d1c442749c2a436406d952107d7920fd0d Dec 03 13:30:52 crc kubenswrapper[4690]: W1203 13:30:52.251130 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ca950cf_73a9_47ae_b07c_b8806057e24b.slice/crio-7ab6d0578410c7caf5841c66eb839af625156f9a0bd5bc5ddce80061bded6529 WatchSource:0}: Error finding container 7ab6d0578410c7caf5841c66eb839af625156f9a0bd5bc5ddce80061bded6529: Status 404 returned error can't find the container with id 7ab6d0578410c7caf5841c66eb839af625156f9a0bd5bc5ddce80061bded6529 Dec 03 13:30:52 crc kubenswrapper[4690]: W1203 13:30:52.252686 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaac7e979_c496_4a3a_8cf0_d3a34d659b85.slice/crio-22c808748ec55aca1f0f35448a82acea16d3da732b6e89e17412bc7f40e7c72a WatchSource:0}: Error finding container 22c808748ec55aca1f0f35448a82acea16d3da732b6e89e17412bc7f40e7c72a: Status 404 returned error can't find the container with id 22c808748ec55aca1f0f35448a82acea16d3da732b6e89e17412bc7f40e7c72a Dec 03 13:30:52 crc kubenswrapper[4690]: I1203 13:30:52.332166 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ce1db85-6b25-47fd-84ee-47406aa77eec" path="/var/lib/kubelet/pods/4ce1db85-6b25-47fd-84ee-47406aa77eec/volumes" Dec 03 13:30:52 crc kubenswrapper[4690]: I1203 13:30:52.516412 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 13:30:52 crc kubenswrapper[4690]: I1203 13:30:52.831805 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rf676"] Dec 03 13:30:52 crc kubenswrapper[4690]: I1203 13:30:52.843456 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-deae-account-create-update-49dp2"] Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.216248 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rf676" event={"ID":"9c88c34d-c55a-4456-9f8b-daf83b54896a","Type":"ContainerStarted","Data":"a20e14f317c449e34abb455815a00507875b98e9afbc46c3683065de31818ec7"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.216894 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rf676" event={"ID":"9c88c34d-c55a-4456-9f8b-daf83b54896a","Type":"ContainerStarted","Data":"aca4caafa2fd14fb634c94139a04cef457b131c8b15b43b658774d042d09a558"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.218386 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wn8pw" event={"ID":"6bb598ea-0a2a-49b6-a989-533ab58cd524","Type":"ContainerStarted","Data":"ec874ce6809ef9799dbfd91855c47d4be6871ade2e1f1816dc7f3a10f915a9a0"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.218429 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wn8pw" event={"ID":"6bb598ea-0a2a-49b6-a989-533ab58cd524","Type":"ContainerStarted","Data":"885e59b01771d1e4ee3f8e7b108370d1c442749c2a436406d952107d7920fd0d"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.220110 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8784-account-create-update-nc4fz" event={"ID":"9cf94540-efe2-4e03-80ff-05dbef2b7566","Type":"ContainerStarted","Data":"d5a560113cb0c1ddb073b28d1c0ebdceb9d951d0f23e024af91b5054e49a45d8"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.220167 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8784-account-create-update-nc4fz" event={"ID":"9cf94540-efe2-4e03-80ff-05dbef2b7566","Type":"ContainerStarted","Data":"a88e5e017aa0749291406f813e4389187c846a77e13808e6d3d8ab5a2abccfde"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.221885 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a217-account-create-update-9pg4q" event={"ID":"aac7e979-c496-4a3a-8cf0-d3a34d659b85","Type":"ContainerStarted","Data":"a1f5b9092d0f7088b1f4e64d7af42c003768d24cba564cbea668ffb8a107dd98"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.221923 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a217-account-create-update-9pg4q" event={"ID":"aac7e979-c496-4a3a-8cf0-d3a34d659b85","Type":"ContainerStarted","Data":"22c808748ec55aca1f0f35448a82acea16d3da732b6e89e17412bc7f40e7c72a"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.224041 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-deae-account-create-update-49dp2" event={"ID":"54bfb9d7-bab2-4541-abd5-96f850bffdde","Type":"ContainerStarted","Data":"33d38665059fcd2b25d5b957f41faa322872860d1668172ae577b70954264b5a"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.224103 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-deae-account-create-update-49dp2" event={"ID":"54bfb9d7-bab2-4541-abd5-96f850bffdde","Type":"ContainerStarted","Data":"8bb62254de2b7c6c7203f04a1380200abac583b87e097a4c32b398d41aadad11"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.226671 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5bb56" event={"ID":"3ca950cf-73a9-47ae-b07c-b8806057e24b","Type":"ContainerStarted","Data":"a7afb4538e398dd6c406596836c6ce88468bd7d7a96425af59939745d0c9e93d"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.226707 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5bb56" event={"ID":"3ca950cf-73a9-47ae-b07c-b8806057e24b","Type":"ContainerStarted","Data":"7ab6d0578410c7caf5841c66eb839af625156f9a0bd5bc5ddce80061bded6529"} Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.238515 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-rf676" podStartSLOduration=2.238495007 podStartE2EDuration="2.238495007s" podCreationTimestamp="2025-12-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:30:53.235805855 +0000 UTC m=+1299.216726288" watchObservedRunningTime="2025-12-03 13:30:53.238495007 +0000 UTC m=+1299.219415440" Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.259240 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-deae-account-create-update-49dp2" podStartSLOduration=2.25921485 podStartE2EDuration="2.25921485s" podCreationTimestamp="2025-12-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:30:53.254552102 +0000 UTC m=+1299.235472535" watchObservedRunningTime="2025-12-03 13:30:53.25921485 +0000 UTC m=+1299.240135283" Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.283330 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8784-account-create-update-nc4fz" podStartSLOduration=3.283304662 podStartE2EDuration="3.283304662s" podCreationTimestamp="2025-12-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:30:53.275620933 +0000 UTC m=+1299.256541376" watchObservedRunningTime="2025-12-03 13:30:53.283304662 +0000 UTC m=+1299.264225095" Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.292956 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-a217-account-create-update-9pg4q" podStartSLOduration=3.292935536 podStartE2EDuration="3.292935536s" podCreationTimestamp="2025-12-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:30:53.289451315 +0000 UTC m=+1299.270371768" watchObservedRunningTime="2025-12-03 13:30:53.292935536 +0000 UTC m=+1299.273855969" Dec 03 13:30:53 crc kubenswrapper[4690]: I1203 13:30:53.308823 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-5bb56" podStartSLOduration=3.308803506 podStartE2EDuration="3.308803506s" podCreationTimestamp="2025-12-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:30:53.30256368 +0000 UTC m=+1299.283484123" watchObservedRunningTime="2025-12-03 13:30:53.308803506 +0000 UTC m=+1299.289723939" Dec 03 13:30:54 crc kubenswrapper[4690]: I1203 13:30:54.238262 4690 generic.go:334] "Generic (PLEG): container finished" podID="9c88c34d-c55a-4456-9f8b-daf83b54896a" containerID="a20e14f317c449e34abb455815a00507875b98e9afbc46c3683065de31818ec7" exitCode=0 Dec 03 13:30:54 crc kubenswrapper[4690]: I1203 13:30:54.238349 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rf676" event={"ID":"9c88c34d-c55a-4456-9f8b-daf83b54896a","Type":"ContainerDied","Data":"a20e14f317c449e34abb455815a00507875b98e9afbc46c3683065de31818ec7"} Dec 03 13:30:54 crc kubenswrapper[4690]: I1203 13:30:54.242765 4690 generic.go:334] "Generic (PLEG): container finished" podID="6bb598ea-0a2a-49b6-a989-533ab58cd524" containerID="ec874ce6809ef9799dbfd91855c47d4be6871ade2e1f1816dc7f3a10f915a9a0" exitCode=0 Dec 03 13:30:54 crc kubenswrapper[4690]: I1203 13:30:54.242829 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wn8pw" event={"ID":"6bb598ea-0a2a-49b6-a989-533ab58cd524","Type":"ContainerDied","Data":"ec874ce6809ef9799dbfd91855c47d4be6871ade2e1f1816dc7f3a10f915a9a0"} Dec 03 13:30:54 crc kubenswrapper[4690]: I1203 13:30:54.245638 4690 generic.go:334] "Generic (PLEG): container finished" podID="3ca950cf-73a9-47ae-b07c-b8806057e24b" containerID="a7afb4538e398dd6c406596836c6ce88468bd7d7a96425af59939745d0c9e93d" exitCode=0 Dec 03 13:30:54 crc kubenswrapper[4690]: I1203 13:30:54.245753 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5bb56" event={"ID":"3ca950cf-73a9-47ae-b07c-b8806057e24b","Type":"ContainerDied","Data":"a7afb4538e398dd6c406596836c6ce88468bd7d7a96425af59939745d0c9e93d"} Dec 03 13:30:54 crc kubenswrapper[4690]: I1203 13:30:54.263399 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-wn8pw" podStartSLOduration=4.26337753 podStartE2EDuration="4.26337753s" podCreationTimestamp="2025-12-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:30:53.317751644 +0000 UTC m=+1299.298672087" watchObservedRunningTime="2025-12-03 13:30:54.26337753 +0000 UTC m=+1300.244297973" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.255662 4690 generic.go:334] "Generic (PLEG): container finished" podID="aac7e979-c496-4a3a-8cf0-d3a34d659b85" containerID="a1f5b9092d0f7088b1f4e64d7af42c003768d24cba564cbea668ffb8a107dd98" exitCode=0 Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.255765 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a217-account-create-update-9pg4q" event={"ID":"aac7e979-c496-4a3a-8cf0-d3a34d659b85","Type":"ContainerDied","Data":"a1f5b9092d0f7088b1f4e64d7af42c003768d24cba564cbea668ffb8a107dd98"} Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.258752 4690 generic.go:334] "Generic (PLEG): container finished" podID="262bf571-0151-4c11-a05c-1270656144f9" containerID="9a04bd0b28e1da640266c690eeca3c69b4ef7a54cb18526c3d6aae705a4eec47" exitCode=0 Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.258816 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bl55c" event={"ID":"262bf571-0151-4c11-a05c-1270656144f9","Type":"ContainerDied","Data":"9a04bd0b28e1da640266c690eeca3c69b4ef7a54cb18526c3d6aae705a4eec47"} Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.703524 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5bb56" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.841893 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.847982 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rf676" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.866691 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ca950cf-73a9-47ae-b07c-b8806057e24b-operator-scripts\") pod \"3ca950cf-73a9-47ae-b07c-b8806057e24b\" (UID: \"3ca950cf-73a9-47ae-b07c-b8806057e24b\") " Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.866813 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cftf9\" (UniqueName: \"kubernetes.io/projected/3ca950cf-73a9-47ae-b07c-b8806057e24b-kube-api-access-cftf9\") pod \"3ca950cf-73a9-47ae-b07c-b8806057e24b\" (UID: \"3ca950cf-73a9-47ae-b07c-b8806057e24b\") " Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.872791 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ca950cf-73a9-47ae-b07c-b8806057e24b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3ca950cf-73a9-47ae-b07c-b8806057e24b" (UID: "3ca950cf-73a9-47ae-b07c-b8806057e24b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.877511 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca950cf-73a9-47ae-b07c-b8806057e24b-kube-api-access-cftf9" (OuterVolumeSpecName: "kube-api-access-cftf9") pod "3ca950cf-73a9-47ae-b07c-b8806057e24b" (UID: "3ca950cf-73a9-47ae-b07c-b8806057e24b"). InnerVolumeSpecName "kube-api-access-cftf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.968406 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz8vc\" (UniqueName: \"kubernetes.io/projected/9c88c34d-c55a-4456-9f8b-daf83b54896a-kube-api-access-dz8vc\") pod \"9c88c34d-c55a-4456-9f8b-daf83b54896a\" (UID: \"9c88c34d-c55a-4456-9f8b-daf83b54896a\") " Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.968548 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb598ea-0a2a-49b6-a989-533ab58cd524-operator-scripts\") pod \"6bb598ea-0a2a-49b6-a989-533ab58cd524\" (UID: \"6bb598ea-0a2a-49b6-a989-533ab58cd524\") " Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.968794 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c88c34d-c55a-4456-9f8b-daf83b54896a-operator-scripts\") pod \"9c88c34d-c55a-4456-9f8b-daf83b54896a\" (UID: \"9c88c34d-c55a-4456-9f8b-daf83b54896a\") " Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.968822 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssl2g\" (UniqueName: \"kubernetes.io/projected/6bb598ea-0a2a-49b6-a989-533ab58cd524-kube-api-access-ssl2g\") pod \"6bb598ea-0a2a-49b6-a989-533ab58cd524\" (UID: \"6bb598ea-0a2a-49b6-a989-533ab58cd524\") " Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.969267 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ca950cf-73a9-47ae-b07c-b8806057e24b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.969294 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cftf9\" (UniqueName: \"kubernetes.io/projected/3ca950cf-73a9-47ae-b07c-b8806057e24b-kube-api-access-cftf9\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.969123 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bb598ea-0a2a-49b6-a989-533ab58cd524-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6bb598ea-0a2a-49b6-a989-533ab58cd524" (UID: "6bb598ea-0a2a-49b6-a989-533ab58cd524"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.969345 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c88c34d-c55a-4456-9f8b-daf83b54896a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c88c34d-c55a-4456-9f8b-daf83b54896a" (UID: "9c88c34d-c55a-4456-9f8b-daf83b54896a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.972042 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bb598ea-0a2a-49b6-a989-533ab58cd524-kube-api-access-ssl2g" (OuterVolumeSpecName: "kube-api-access-ssl2g") pod "6bb598ea-0a2a-49b6-a989-533ab58cd524" (UID: "6bb598ea-0a2a-49b6-a989-533ab58cd524"). InnerVolumeSpecName "kube-api-access-ssl2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:55 crc kubenswrapper[4690]: I1203 13:30:55.972580 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c88c34d-c55a-4456-9f8b-daf83b54896a-kube-api-access-dz8vc" (OuterVolumeSpecName: "kube-api-access-dz8vc") pod "9c88c34d-c55a-4456-9f8b-daf83b54896a" (UID: "9c88c34d-c55a-4456-9f8b-daf83b54896a"). InnerVolumeSpecName "kube-api-access-dz8vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.071034 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c88c34d-c55a-4456-9f8b-daf83b54896a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.071088 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssl2g\" (UniqueName: \"kubernetes.io/projected/6bb598ea-0a2a-49b6-a989-533ab58cd524-kube-api-access-ssl2g\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.071104 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz8vc\" (UniqueName: \"kubernetes.io/projected/9c88c34d-c55a-4456-9f8b-daf83b54896a-kube-api-access-dz8vc\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.071118 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb598ea-0a2a-49b6-a989-533ab58cd524-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.273093 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rf676" event={"ID":"9c88c34d-c55a-4456-9f8b-daf83b54896a","Type":"ContainerDied","Data":"aca4caafa2fd14fb634c94139a04cef457b131c8b15b43b658774d042d09a558"} Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.273157 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aca4caafa2fd14fb634c94139a04cef457b131c8b15b43b658774d042d09a558" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.273132 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rf676" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.275673 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wn8pw" event={"ID":"6bb598ea-0a2a-49b6-a989-533ab58cd524","Type":"ContainerDied","Data":"885e59b01771d1e4ee3f8e7b108370d1c442749c2a436406d952107d7920fd0d"} Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.275770 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="885e59b01771d1e4ee3f8e7b108370d1c442749c2a436406d952107d7920fd0d" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.275793 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wn8pw" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.280625 4690 generic.go:334] "Generic (PLEG): container finished" podID="9cf94540-efe2-4e03-80ff-05dbef2b7566" containerID="d5a560113cb0c1ddb073b28d1c0ebdceb9d951d0f23e024af91b5054e49a45d8" exitCode=0 Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.280711 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8784-account-create-update-nc4fz" event={"ID":"9cf94540-efe2-4e03-80ff-05dbef2b7566","Type":"ContainerDied","Data":"d5a560113cb0c1ddb073b28d1c0ebdceb9d951d0f23e024af91b5054e49a45d8"} Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.282909 4690 generic.go:334] "Generic (PLEG): container finished" podID="54bfb9d7-bab2-4541-abd5-96f850bffdde" containerID="33d38665059fcd2b25d5b957f41faa322872860d1668172ae577b70954264b5a" exitCode=0 Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.282983 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-deae-account-create-update-49dp2" event={"ID":"54bfb9d7-bab2-4541-abd5-96f850bffdde","Type":"ContainerDied","Data":"33d38665059fcd2b25d5b957f41faa322872860d1668172ae577b70954264b5a"} Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.289196 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5bb56" event={"ID":"3ca950cf-73a9-47ae-b07c-b8806057e24b","Type":"ContainerDied","Data":"7ab6d0578410c7caf5841c66eb839af625156f9a0bd5bc5ddce80061bded6529"} Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.289233 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5bb56" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.289252 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ab6d0578410c7caf5841c66eb839af625156f9a0bd5bc5ddce80061bded6529" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.836575 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.852539 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.987772 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-scripts\") pod \"262bf571-0151-4c11-a05c-1270656144f9\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.987910 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-swiftconf\") pod \"262bf571-0151-4c11-a05c-1270656144f9\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.987962 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-dispersionconf\") pod \"262bf571-0151-4c11-a05c-1270656144f9\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.987986 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rrx6\" (UniqueName: \"kubernetes.io/projected/aac7e979-c496-4a3a-8cf0-d3a34d659b85-kube-api-access-6rrx6\") pod \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\" (UID: \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.988039 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-ring-data-devices\") pod \"262bf571-0151-4c11-a05c-1270656144f9\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.988085 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/262bf571-0151-4c11-a05c-1270656144f9-etc-swift\") pod \"262bf571-0151-4c11-a05c-1270656144f9\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.988947 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwt2b\" (UniqueName: \"kubernetes.io/projected/262bf571-0151-4c11-a05c-1270656144f9-kube-api-access-xwt2b\") pod \"262bf571-0151-4c11-a05c-1270656144f9\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.989075 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac7e979-c496-4a3a-8cf0-d3a34d659b85-operator-scripts\") pod \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\" (UID: \"aac7e979-c496-4a3a-8cf0-d3a34d659b85\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.989137 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-combined-ca-bundle\") pod \"262bf571-0151-4c11-a05c-1270656144f9\" (UID: \"262bf571-0151-4c11-a05c-1270656144f9\") " Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.988958 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "262bf571-0151-4c11-a05c-1270656144f9" (UID: "262bf571-0151-4c11-a05c-1270656144f9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.989385 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/262bf571-0151-4c11-a05c-1270656144f9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "262bf571-0151-4c11-a05c-1270656144f9" (UID: "262bf571-0151-4c11-a05c-1270656144f9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.989482 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aac7e979-c496-4a3a-8cf0-d3a34d659b85-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aac7e979-c496-4a3a-8cf0-d3a34d659b85" (UID: "aac7e979-c496-4a3a-8cf0-d3a34d659b85"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.990177 4690 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.990206 4690 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/262bf571-0151-4c11-a05c-1270656144f9-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.990218 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac7e979-c496-4a3a-8cf0-d3a34d659b85-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.993863 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/262bf571-0151-4c11-a05c-1270656144f9-kube-api-access-xwt2b" (OuterVolumeSpecName: "kube-api-access-xwt2b") pod "262bf571-0151-4c11-a05c-1270656144f9" (UID: "262bf571-0151-4c11-a05c-1270656144f9"). InnerVolumeSpecName "kube-api-access-xwt2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.995894 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac7e979-c496-4a3a-8cf0-d3a34d659b85-kube-api-access-6rrx6" (OuterVolumeSpecName: "kube-api-access-6rrx6") pod "aac7e979-c496-4a3a-8cf0-d3a34d659b85" (UID: "aac7e979-c496-4a3a-8cf0-d3a34d659b85"). InnerVolumeSpecName "kube-api-access-6rrx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:56 crc kubenswrapper[4690]: I1203 13:30:56.996102 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "262bf571-0151-4c11-a05c-1270656144f9" (UID: "262bf571-0151-4c11-a05c-1270656144f9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.012649 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-scripts" (OuterVolumeSpecName: "scripts") pod "262bf571-0151-4c11-a05c-1270656144f9" (UID: "262bf571-0151-4c11-a05c-1270656144f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.017734 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "262bf571-0151-4c11-a05c-1270656144f9" (UID: "262bf571-0151-4c11-a05c-1270656144f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.019897 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "262bf571-0151-4c11-a05c-1270656144f9" (UID: "262bf571-0151-4c11-a05c-1270656144f9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.091654 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/262bf571-0151-4c11-a05c-1270656144f9-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.091693 4690 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.091704 4690 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.091715 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rrx6\" (UniqueName: \"kubernetes.io/projected/aac7e979-c496-4a3a-8cf0-d3a34d659b85-kube-api-access-6rrx6\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.091725 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwt2b\" (UniqueName: \"kubernetes.io/projected/262bf571-0151-4c11-a05c-1270656144f9-kube-api-access-xwt2b\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.091733 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262bf571-0151-4c11-a05c-1270656144f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.300293 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a217-account-create-update-9pg4q" event={"ID":"aac7e979-c496-4a3a-8cf0-d3a34d659b85","Type":"ContainerDied","Data":"22c808748ec55aca1f0f35448a82acea16d3da732b6e89e17412bc7f40e7c72a"} Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.300388 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22c808748ec55aca1f0f35448a82acea16d3da732b6e89e17412bc7f40e7c72a" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.300323 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a217-account-create-update-9pg4q" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.302545 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bl55c" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.304029 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bl55c" event={"ID":"262bf571-0151-4c11-a05c-1270656144f9","Type":"ContainerDied","Data":"19eea06d789dd49925ccf86d413227fcec4a4602763d473be48e370089f1a8ba"} Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.304108 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19eea06d789dd49925ccf86d413227fcec4a4602763d473be48e370089f1a8ba" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.447900 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.460149 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-znnxj" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.625376 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-mf8xw" podUID="64710eed-cdda-4f62-a007-766dd92f0b29" containerName="ovn-controller" probeResult="failure" output=< Dec 03 13:30:57 crc kubenswrapper[4690]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 13:30:57 crc kubenswrapper[4690]: > Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.759890 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mf8xw-config-pkgqp"] Dec 03 13:30:57 crc kubenswrapper[4690]: E1203 13:30:57.760416 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb598ea-0a2a-49b6-a989-533ab58cd524" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760441 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb598ea-0a2a-49b6-a989-533ab58cd524" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: E1203 13:30:57.760460 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca950cf-73a9-47ae-b07c-b8806057e24b" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760468 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca950cf-73a9-47ae-b07c-b8806057e24b" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: E1203 13:30:57.760484 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac7e979-c496-4a3a-8cf0-d3a34d659b85" containerName="mariadb-account-create-update" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760492 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac7e979-c496-4a3a-8cf0-d3a34d659b85" containerName="mariadb-account-create-update" Dec 03 13:30:57 crc kubenswrapper[4690]: E1203 13:30:57.760512 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c88c34d-c55a-4456-9f8b-daf83b54896a" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760520 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c88c34d-c55a-4456-9f8b-daf83b54896a" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: E1203 13:30:57.760543 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262bf571-0151-4c11-a05c-1270656144f9" containerName="swift-ring-rebalance" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760552 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="262bf571-0151-4c11-a05c-1270656144f9" containerName="swift-ring-rebalance" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760775 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bb598ea-0a2a-49b6-a989-533ab58cd524" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760788 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="262bf571-0151-4c11-a05c-1270656144f9" containerName="swift-ring-rebalance" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760801 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca950cf-73a9-47ae-b07c-b8806057e24b" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760824 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac7e979-c496-4a3a-8cf0-d3a34d659b85" containerName="mariadb-account-create-update" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.760834 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c88c34d-c55a-4456-9f8b-daf83b54896a" containerName="mariadb-database-create" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.761703 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.764272 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.769094 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mf8xw-config-pkgqp"] Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.907501 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-log-ovn\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.907561 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-additional-scripts\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.907630 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run-ovn\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.907737 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4kzx\" (UniqueName: \"kubernetes.io/projected/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-kube-api-access-n4kzx\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.907785 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.907881 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-scripts\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.917073 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:57 crc kubenswrapper[4690]: I1203 13:30:57.924233 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.009589 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kdsm\" (UniqueName: \"kubernetes.io/projected/54bfb9d7-bab2-4541-abd5-96f850bffdde-kube-api-access-5kdsm\") pod \"54bfb9d7-bab2-4541-abd5-96f850bffdde\" (UID: \"54bfb9d7-bab2-4541-abd5-96f850bffdde\") " Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.009732 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bfb9d7-bab2-4541-abd5-96f850bffdde-operator-scripts\") pod \"54bfb9d7-bab2-4541-abd5-96f850bffdde\" (UID: \"54bfb9d7-bab2-4541-abd5-96f850bffdde\") " Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010082 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4kzx\" (UniqueName: \"kubernetes.io/projected/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-kube-api-access-n4kzx\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010151 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010205 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-scripts\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010629 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-log-ovn\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010697 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-additional-scripts\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010705 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-log-ovn\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010638 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54bfb9d7-bab2-4541-abd5-96f850bffdde-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54bfb9d7-bab2-4541-abd5-96f850bffdde" (UID: "54bfb9d7-bab2-4541-abd5-96f850bffdde"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010706 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010838 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run-ovn\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.010966 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bfb9d7-bab2-4541-abd5-96f850bffdde-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.011061 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run-ovn\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.011378 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-additional-scripts\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.012557 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-scripts\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.015307 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54bfb9d7-bab2-4541-abd5-96f850bffdde-kube-api-access-5kdsm" (OuterVolumeSpecName: "kube-api-access-5kdsm") pod "54bfb9d7-bab2-4541-abd5-96f850bffdde" (UID: "54bfb9d7-bab2-4541-abd5-96f850bffdde"). InnerVolumeSpecName "kube-api-access-5kdsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.030754 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4kzx\" (UniqueName: \"kubernetes.io/projected/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-kube-api-access-n4kzx\") pod \"ovn-controller-mf8xw-config-pkgqp\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.084695 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.112811 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf94540-efe2-4e03-80ff-05dbef2b7566-operator-scripts\") pod \"9cf94540-efe2-4e03-80ff-05dbef2b7566\" (UID: \"9cf94540-efe2-4e03-80ff-05dbef2b7566\") " Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.113480 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xnsf\" (UniqueName: \"kubernetes.io/projected/9cf94540-efe2-4e03-80ff-05dbef2b7566-kube-api-access-7xnsf\") pod \"9cf94540-efe2-4e03-80ff-05dbef2b7566\" (UID: \"9cf94540-efe2-4e03-80ff-05dbef2b7566\") " Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.113482 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cf94540-efe2-4e03-80ff-05dbef2b7566-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9cf94540-efe2-4e03-80ff-05dbef2b7566" (UID: "9cf94540-efe2-4e03-80ff-05dbef2b7566"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.114697 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf94540-efe2-4e03-80ff-05dbef2b7566-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.114960 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kdsm\" (UniqueName: \"kubernetes.io/projected/54bfb9d7-bab2-4541-abd5-96f850bffdde-kube-api-access-5kdsm\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.117850 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cf94540-efe2-4e03-80ff-05dbef2b7566-kube-api-access-7xnsf" (OuterVolumeSpecName: "kube-api-access-7xnsf") pod "9cf94540-efe2-4e03-80ff-05dbef2b7566" (UID: "9cf94540-efe2-4e03-80ff-05dbef2b7566"). InnerVolumeSpecName "kube-api-access-7xnsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.217350 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xnsf\" (UniqueName: \"kubernetes.io/projected/9cf94540-efe2-4e03-80ff-05dbef2b7566-kube-api-access-7xnsf\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.338115 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8784-account-create-update-nc4fz" event={"ID":"9cf94540-efe2-4e03-80ff-05dbef2b7566","Type":"ContainerDied","Data":"a88e5e017aa0749291406f813e4389187c846a77e13808e6d3d8ab5a2abccfde"} Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.339601 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a88e5e017aa0749291406f813e4389187c846a77e13808e6d3d8ab5a2abccfde" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.338410 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8784-account-create-update-nc4fz" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.350159 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-deae-account-create-update-49dp2" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.350164 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-deae-account-create-update-49dp2" event={"ID":"54bfb9d7-bab2-4541-abd5-96f850bffdde","Type":"ContainerDied","Data":"8bb62254de2b7c6c7203f04a1380200abac583b87e097a4c32b398d41aadad11"} Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.350287 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bb62254de2b7c6c7203f04a1380200abac583b87e097a4c32b398d41aadad11" Dec 03 13:30:58 crc kubenswrapper[4690]: I1203 13:30:58.590020 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mf8xw-config-pkgqp"] Dec 03 13:30:58 crc kubenswrapper[4690]: W1203 13:30:58.590842 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f496b7b_7a93_4f6f_a25d_67a7473d89e6.slice/crio-83ef3b1d6ccd44aa5161fb5e03cf04fc92dcc8e945e1a94ca53d897899028fa4 WatchSource:0}: Error finding container 83ef3b1d6ccd44aa5161fb5e03cf04fc92dcc8e945e1a94ca53d897899028fa4: Status 404 returned error can't find the container with id 83ef3b1d6ccd44aa5161fb5e03cf04fc92dcc8e945e1a94ca53d897899028fa4 Dec 03 13:30:59 crc kubenswrapper[4690]: I1203 13:30:59.362346 4690 generic.go:334] "Generic (PLEG): container finished" podID="6173a433-90f9-4e48-8297-81767d5ce03e" containerID="ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670" exitCode=0 Dec 03 13:30:59 crc kubenswrapper[4690]: I1203 13:30:59.362438 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6173a433-90f9-4e48-8297-81767d5ce03e","Type":"ContainerDied","Data":"ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670"} Dec 03 13:30:59 crc kubenswrapper[4690]: I1203 13:30:59.367357 4690 generic.go:334] "Generic (PLEG): container finished" podID="3f496b7b-7a93-4f6f-a25d-67a7473d89e6" containerID="6aa3703018330b11655b823fe6b157d787f28a70f379fdebf89ef23a9401accf" exitCode=0 Dec 03 13:30:59 crc kubenswrapper[4690]: I1203 13:30:59.367433 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mf8xw-config-pkgqp" event={"ID":"3f496b7b-7a93-4f6f-a25d-67a7473d89e6","Type":"ContainerDied","Data":"6aa3703018330b11655b823fe6b157d787f28a70f379fdebf89ef23a9401accf"} Dec 03 13:30:59 crc kubenswrapper[4690]: I1203 13:30:59.367477 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mf8xw-config-pkgqp" event={"ID":"3f496b7b-7a93-4f6f-a25d-67a7473d89e6","Type":"ContainerStarted","Data":"83ef3b1d6ccd44aa5161fb5e03cf04fc92dcc8e945e1a94ca53d897899028fa4"} Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.379091 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6173a433-90f9-4e48-8297-81767d5ce03e","Type":"ContainerStarted","Data":"e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1"} Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.379446 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.380315 4690 generic.go:334] "Generic (PLEG): container finished" podID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerID="4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928" exitCode=0 Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.380498 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5930069-12cd-4b0b-8aab-0f32ea9057a9","Type":"ContainerDied","Data":"4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928"} Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.416745 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=46.671820985 podStartE2EDuration="1m24.416722701s" podCreationTimestamp="2025-12-03 13:29:36 +0000 UTC" firstStartedPulling="2025-12-03 13:29:45.887296962 +0000 UTC m=+1231.868217395" lastFinishedPulling="2025-12-03 13:30:23.632198688 +0000 UTC m=+1269.613119111" observedRunningTime="2025-12-03 13:31:00.405353847 +0000 UTC m=+1306.386274300" watchObservedRunningTime="2025-12-03 13:31:00.416722701 +0000 UTC m=+1306.397643134" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.756163 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.875093 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-scripts\") pod \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.875171 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4kzx\" (UniqueName: \"kubernetes.io/projected/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-kube-api-access-n4kzx\") pod \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.875206 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-additional-scripts\") pod \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.875252 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run\") pod \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.875355 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run-ovn\") pod \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.875589 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-log-ovn\") pod \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\" (UID: \"3f496b7b-7a93-4f6f-a25d-67a7473d89e6\") " Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.876149 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "3f496b7b-7a93-4f6f-a25d-67a7473d89e6" (UID: "3f496b7b-7a93-4f6f-a25d-67a7473d89e6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.876535 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-scripts" (OuterVolumeSpecName: "scripts") pod "3f496b7b-7a93-4f6f-a25d-67a7473d89e6" (UID: "3f496b7b-7a93-4f6f-a25d-67a7473d89e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.876619 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run" (OuterVolumeSpecName: "var-run") pod "3f496b7b-7a93-4f6f-a25d-67a7473d89e6" (UID: "3f496b7b-7a93-4f6f-a25d-67a7473d89e6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.876641 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "3f496b7b-7a93-4f6f-a25d-67a7473d89e6" (UID: "3f496b7b-7a93-4f6f-a25d-67a7473d89e6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.877140 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "3f496b7b-7a93-4f6f-a25d-67a7473d89e6" (UID: "3f496b7b-7a93-4f6f-a25d-67a7473d89e6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.880359 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-kube-api-access-n4kzx" (OuterVolumeSpecName: "kube-api-access-n4kzx") pod "3f496b7b-7a93-4f6f-a25d-67a7473d89e6" (UID: "3f496b7b-7a93-4f6f-a25d-67a7473d89e6"). InnerVolumeSpecName "kube-api-access-n4kzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.978091 4690 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.978144 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.978155 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4kzx\" (UniqueName: \"kubernetes.io/projected/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-kube-api-access-n4kzx\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.978165 4690 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.978175 4690 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:00 crc kubenswrapper[4690]: I1203 13:31:00.978182 4690 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3f496b7b-7a93-4f6f-a25d-67a7473d89e6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.391676 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5930069-12cd-4b0b-8aab-0f32ea9057a9","Type":"ContainerStarted","Data":"65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807"} Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.391910 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.394010 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mf8xw-config-pkgqp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.400686 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mf8xw-config-pkgqp" event={"ID":"3f496b7b-7a93-4f6f-a25d-67a7473d89e6","Type":"ContainerDied","Data":"83ef3b1d6ccd44aa5161fb5e03cf04fc92dcc8e945e1a94ca53d897899028fa4"} Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.400746 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83ef3b1d6ccd44aa5161fb5e03cf04fc92dcc8e945e1a94ca53d897899028fa4" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.441962 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371951.412842 podStartE2EDuration="1m25.441933442s" podCreationTimestamp="2025-12-03 13:29:36 +0000 UTC" firstStartedPulling="2025-12-03 13:29:45.887711492 +0000 UTC m=+1231.868631925" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:01.43629991 +0000 UTC m=+1307.417220353" watchObservedRunningTime="2025-12-03 13:31:01.441933442 +0000 UTC m=+1307.422853875" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.485186 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-fq5zp"] Dec 03 13:31:01 crc kubenswrapper[4690]: E1203 13:31:01.485695 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f496b7b-7a93-4f6f-a25d-67a7473d89e6" containerName="ovn-config" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.485726 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f496b7b-7a93-4f6f-a25d-67a7473d89e6" containerName="ovn-config" Dec 03 13:31:01 crc kubenswrapper[4690]: E1203 13:31:01.485748 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf94540-efe2-4e03-80ff-05dbef2b7566" containerName="mariadb-account-create-update" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.485759 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf94540-efe2-4e03-80ff-05dbef2b7566" containerName="mariadb-account-create-update" Dec 03 13:31:01 crc kubenswrapper[4690]: E1203 13:31:01.485785 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54bfb9d7-bab2-4541-abd5-96f850bffdde" containerName="mariadb-account-create-update" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.485793 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="54bfb9d7-bab2-4541-abd5-96f850bffdde" containerName="mariadb-account-create-update" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.486083 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f496b7b-7a93-4f6f-a25d-67a7473d89e6" containerName="ovn-config" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.486123 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="54bfb9d7-bab2-4541-abd5-96f850bffdde" containerName="mariadb-account-create-update" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.486142 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cf94540-efe2-4e03-80ff-05dbef2b7566" containerName="mariadb-account-create-update" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.488495 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.490902 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ld4lm" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.490986 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.501843 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fq5zp"] Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.589106 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-db-sync-config-data\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.589189 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-config-data\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.589281 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7vd\" (UniqueName: \"kubernetes.io/projected/f1a38ea3-0bc0-4533-bbec-de05b19433c7-kube-api-access-dh7vd\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.589350 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-combined-ca-bundle\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.691758 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-config-data\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.692642 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7vd\" (UniqueName: \"kubernetes.io/projected/f1a38ea3-0bc0-4533-bbec-de05b19433c7-kube-api-access-dh7vd\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.692816 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-combined-ca-bundle\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.692988 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-db-sync-config-data\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.698176 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-combined-ca-bundle\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.698645 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-db-sync-config-data\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.700247 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-config-data\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.714176 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7vd\" (UniqueName: \"kubernetes.io/projected/f1a38ea3-0bc0-4533-bbec-de05b19433c7-kube-api-access-dh7vd\") pod \"glance-db-sync-fq5zp\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.806501 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fq5zp" Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.908964 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-mf8xw-config-pkgqp"] Dec 03 13:31:01 crc kubenswrapper[4690]: I1203 13:31:01.919359 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-mf8xw-config-pkgqp"] Dec 03 13:31:02 crc kubenswrapper[4690]: I1203 13:31:02.325708 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f496b7b-7a93-4f6f-a25d-67a7473d89e6" path="/var/lib/kubelet/pods/3f496b7b-7a93-4f6f-a25d-67a7473d89e6/volumes" Dec 03 13:31:04 crc kubenswrapper[4690]: I1203 13:31:02.397432 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-mf8xw" Dec 03 13:31:04 crc kubenswrapper[4690]: I1203 13:31:03.110849 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fq5zp"] Dec 03 13:31:04 crc kubenswrapper[4690]: W1203 13:31:03.115117 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1a38ea3_0bc0_4533_bbec_de05b19433c7.slice/crio-7c53356369419b0fce73a9f5bddd4dcd219016815079ed4075e31d588d2fea33 WatchSource:0}: Error finding container 7c53356369419b0fce73a9f5bddd4dcd219016815079ed4075e31d588d2fea33: Status 404 returned error can't find the container with id 7c53356369419b0fce73a9f5bddd4dcd219016815079ed4075e31d588d2fea33 Dec 03 13:31:04 crc kubenswrapper[4690]: I1203 13:31:03.421669 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fq5zp" event={"ID":"f1a38ea3-0bc0-4533-bbec-de05b19433c7","Type":"ContainerStarted","Data":"7c53356369419b0fce73a9f5bddd4dcd219016815079ed4075e31d588d2fea33"} Dec 03 13:31:06 crc kubenswrapper[4690]: I1203 13:31:06.589692 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:31:06 crc kubenswrapper[4690]: I1203 13:31:06.605766 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28748086-659b-453c-bb91-8d9204374da6-etc-swift\") pod \"swift-storage-0\" (UID: \"28748086-659b-453c-bb91-8d9204374da6\") " pod="openstack/swift-storage-0" Dec 03 13:31:06 crc kubenswrapper[4690]: I1203 13:31:06.701080 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 13:31:08 crc kubenswrapper[4690]: I1203 13:31:08.098763 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 13:31:08 crc kubenswrapper[4690]: W1203 13:31:08.109735 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28748086_659b_453c_bb91_8d9204374da6.slice/crio-79c6944a9e073dbfda3d1fa2c5ff594e97265bc096967f164e78c1fc5246265e WatchSource:0}: Error finding container 79c6944a9e073dbfda3d1fa2c5ff594e97265bc096967f164e78c1fc5246265e: Status 404 returned error can't find the container with id 79c6944a9e073dbfda3d1fa2c5ff594e97265bc096967f164e78c1fc5246265e Dec 03 13:31:08 crc kubenswrapper[4690]: I1203 13:31:08.498522 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"79c6944a9e073dbfda3d1fa2c5ff594e97265bc096967f164e78c1fc5246265e"} Dec 03 13:31:17 crc kubenswrapper[4690]: I1203 13:31:17.890793 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 13:31:17 crc kubenswrapper[4690]: I1203 13:31:17.999167 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:31:18 crc kubenswrapper[4690]: E1203 13:31:18.276098 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 03 13:31:18 crc kubenswrapper[4690]: E1203 13:31:18.276531 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dh7vd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-fq5zp_openstack(f1a38ea3-0bc0-4533-bbec-de05b19433c7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:31:18 crc kubenswrapper[4690]: E1203 13:31:18.277944 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-fq5zp" podUID="f1a38ea3-0bc0-4533-bbec-de05b19433c7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.326369 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-hbltr"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.327488 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.330868 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hbltr"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.427137 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-gx6kj"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.428688 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.439084 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx52v\" (UniqueName: \"kubernetes.io/projected/a58de3c0-aba8-49ce-a59c-67732420e08f-kube-api-access-tx52v\") pod \"cinder-db-create-hbltr\" (UID: \"a58de3c0-aba8-49ce-a59c-67732420e08f\") " pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.439155 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a58de3c0-aba8-49ce-a59c-67732420e08f-operator-scripts\") pod \"cinder-db-create-hbltr\" (UID: \"a58de3c0-aba8-49ce-a59c-67732420e08f\") " pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.444601 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-gx6kj"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.465985 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-dab4-account-create-update-445qb"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.467806 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.473906 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.481157 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dab4-account-create-update-445qb"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.541281 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp4nk\" (UniqueName: \"kubernetes.io/projected/8441b637-1109-48c4-8bd5-efbcd4a34a56-kube-api-access-sp4nk\") pod \"barbican-db-create-gx6kj\" (UID: \"8441b637-1109-48c4-8bd5-efbcd4a34a56\") " pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.541387 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx52v\" (UniqueName: \"kubernetes.io/projected/a58de3c0-aba8-49ce-a59c-67732420e08f-kube-api-access-tx52v\") pod \"cinder-db-create-hbltr\" (UID: \"a58de3c0-aba8-49ce-a59c-67732420e08f\") " pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.541429 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a58de3c0-aba8-49ce-a59c-67732420e08f-operator-scripts\") pod \"cinder-db-create-hbltr\" (UID: \"a58de3c0-aba8-49ce-a59c-67732420e08f\") " pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.541487 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8441b637-1109-48c4-8bd5-efbcd4a34a56-operator-scripts\") pod \"barbican-db-create-gx6kj\" (UID: \"8441b637-1109-48c4-8bd5-efbcd4a34a56\") " pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.542495 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a58de3c0-aba8-49ce-a59c-67732420e08f-operator-scripts\") pod \"cinder-db-create-hbltr\" (UID: \"a58de3c0-aba8-49ce-a59c-67732420e08f\") " pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.553927 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-4898-account-create-update-66xl7"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.567884 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.584372 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.585302 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4898-account-create-update-66xl7"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.599620 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx52v\" (UniqueName: \"kubernetes.io/projected/a58de3c0-aba8-49ce-a59c-67732420e08f-kube-api-access-tx52v\") pod \"cinder-db-create-hbltr\" (UID: \"a58de3c0-aba8-49ce-a59c-67732420e08f\") " pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:18 crc kubenswrapper[4690]: E1203 13:31:18.622158 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-fq5zp" podUID="f1a38ea3-0bc0-4533-bbec-de05b19433c7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.627474 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-cg7p2"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.636514 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.639937 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.641464 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.641657 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wpcrf" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.642207 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.644856 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp4nk\" (UniqueName: \"kubernetes.io/projected/8441b637-1109-48c4-8bd5-efbcd4a34a56-kube-api-access-sp4nk\") pod \"barbican-db-create-gx6kj\" (UID: \"8441b637-1109-48c4-8bd5-efbcd4a34a56\") " pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.645253 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfxnf\" (UniqueName: \"kubernetes.io/projected/4c3e3956-3bd9-4693-9fd1-5040c9f90950-kube-api-access-vfxnf\") pod \"cinder-dab4-account-create-update-445qb\" (UID: \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\") " pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.645371 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-864r8\" (UniqueName: \"kubernetes.io/projected/d7b204d5-9b12-409c-ac51-6a14dc45debe-kube-api-access-864r8\") pod \"barbican-4898-account-create-update-66xl7\" (UID: \"d7b204d5-9b12-409c-ac51-6a14dc45debe\") " pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.645392 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c3e3956-3bd9-4693-9fd1-5040c9f90950-operator-scripts\") pod \"cinder-dab4-account-create-update-445qb\" (UID: \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\") " pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.645410 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8441b637-1109-48c4-8bd5-efbcd4a34a56-operator-scripts\") pod \"barbican-db-create-gx6kj\" (UID: \"8441b637-1109-48c4-8bd5-efbcd4a34a56\") " pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.645436 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b204d5-9b12-409c-ac51-6a14dc45debe-operator-scripts\") pod \"barbican-4898-account-create-update-66xl7\" (UID: \"d7b204d5-9b12-409c-ac51-6a14dc45debe\") " pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.646302 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8441b637-1109-48c4-8bd5-efbcd4a34a56-operator-scripts\") pod \"barbican-db-create-gx6kj\" (UID: \"8441b637-1109-48c4-8bd5-efbcd4a34a56\") " pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.654991 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.682488 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp4nk\" (UniqueName: \"kubernetes.io/projected/8441b637-1109-48c4-8bd5-efbcd4a34a56-kube-api-access-sp4nk\") pod \"barbican-db-create-gx6kj\" (UID: \"8441b637-1109-48c4-8bd5-efbcd4a34a56\") " pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.685644 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-cg7p2"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.694294 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-59mk8"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.695745 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.723052 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-59mk8"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.747202 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-combined-ca-bundle\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.747354 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-config-data\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.747398 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdlmn\" (UniqueName: \"kubernetes.io/projected/153bc447-b9f4-4746-9869-17088a3b71ae-kube-api-access-pdlmn\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.747905 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-864r8\" (UniqueName: \"kubernetes.io/projected/d7b204d5-9b12-409c-ac51-6a14dc45debe-kube-api-access-864r8\") pod \"barbican-4898-account-create-update-66xl7\" (UID: \"d7b204d5-9b12-409c-ac51-6a14dc45debe\") " pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.748005 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c3e3956-3bd9-4693-9fd1-5040c9f90950-operator-scripts\") pod \"cinder-dab4-account-create-update-445qb\" (UID: \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\") " pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.748162 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b204d5-9b12-409c-ac51-6a14dc45debe-operator-scripts\") pod \"barbican-4898-account-create-update-66xl7\" (UID: \"d7b204d5-9b12-409c-ac51-6a14dc45debe\") " pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.748559 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfxnf\" (UniqueName: \"kubernetes.io/projected/4c3e3956-3bd9-4693-9fd1-5040c9f90950-kube-api-access-vfxnf\") pod \"cinder-dab4-account-create-update-445qb\" (UID: \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\") " pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.748884 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c3e3956-3bd9-4693-9fd1-5040c9f90950-operator-scripts\") pod \"cinder-dab4-account-create-update-445qb\" (UID: \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\") " pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.749373 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b204d5-9b12-409c-ac51-6a14dc45debe-operator-scripts\") pod \"barbican-4898-account-create-update-66xl7\" (UID: \"d7b204d5-9b12-409c-ac51-6a14dc45debe\") " pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.751416 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.765536 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfxnf\" (UniqueName: \"kubernetes.io/projected/4c3e3956-3bd9-4693-9fd1-5040c9f90950-kube-api-access-vfxnf\") pod \"cinder-dab4-account-create-update-445qb\" (UID: \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\") " pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.771485 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-864r8\" (UniqueName: \"kubernetes.io/projected/d7b204d5-9b12-409c-ac51-6a14dc45debe-kube-api-access-864r8\") pod \"barbican-4898-account-create-update-66xl7\" (UID: \"d7b204d5-9b12-409c-ac51-6a14dc45debe\") " pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.791296 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.840721 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c6af-account-create-update-bzzjg"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.842054 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.851173 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdlmn\" (UniqueName: \"kubernetes.io/projected/153bc447-b9f4-4746-9869-17088a3b71ae-kube-api-access-pdlmn\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.851373 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-combined-ca-bundle\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.851463 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad027298-b2ff-4145-9a45-2dcbf5b5f228-operator-scripts\") pod \"neutron-db-create-59mk8\" (UID: \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\") " pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.851505 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-config-data\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.851551 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d56pj\" (UniqueName: \"kubernetes.io/projected/ad027298-b2ff-4145-9a45-2dcbf5b5f228-kube-api-access-d56pj\") pod \"neutron-db-create-59mk8\" (UID: \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\") " pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.852056 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.856998 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-config-data\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.857722 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c6af-account-create-update-bzzjg"] Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.863586 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-combined-ca-bundle\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.876373 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdlmn\" (UniqueName: \"kubernetes.io/projected/153bc447-b9f4-4746-9869-17088a3b71ae-kube-api-access-pdlmn\") pod \"keystone-db-sync-cg7p2\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.936189 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.953225 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad027298-b2ff-4145-9a45-2dcbf5b5f228-operator-scripts\") pod \"neutron-db-create-59mk8\" (UID: \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\") " pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.953631 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d56pj\" (UniqueName: \"kubernetes.io/projected/ad027298-b2ff-4145-9a45-2dcbf5b5f228-kube-api-access-d56pj\") pod \"neutron-db-create-59mk8\" (UID: \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\") " pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.953685 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/769615ff-2aa5-41b8-812b-a95a27f72727-operator-scripts\") pod \"neutron-c6af-account-create-update-bzzjg\" (UID: \"769615ff-2aa5-41b8-812b-a95a27f72727\") " pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.953762 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll52c\" (UniqueName: \"kubernetes.io/projected/769615ff-2aa5-41b8-812b-a95a27f72727-kube-api-access-ll52c\") pod \"neutron-c6af-account-create-update-bzzjg\" (UID: \"769615ff-2aa5-41b8-812b-a95a27f72727\") " pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.955551 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad027298-b2ff-4145-9a45-2dcbf5b5f228-operator-scripts\") pod \"neutron-db-create-59mk8\" (UID: \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\") " pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.967090 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:18 crc kubenswrapper[4690]: I1203 13:31:18.973588 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d56pj\" (UniqueName: \"kubernetes.io/projected/ad027298-b2ff-4145-9a45-2dcbf5b5f228-kube-api-access-d56pj\") pod \"neutron-db-create-59mk8\" (UID: \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\") " pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.055688 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/769615ff-2aa5-41b8-812b-a95a27f72727-operator-scripts\") pod \"neutron-c6af-account-create-update-bzzjg\" (UID: \"769615ff-2aa5-41b8-812b-a95a27f72727\") " pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.055789 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll52c\" (UniqueName: \"kubernetes.io/projected/769615ff-2aa5-41b8-812b-a95a27f72727-kube-api-access-ll52c\") pod \"neutron-c6af-account-create-update-bzzjg\" (UID: \"769615ff-2aa5-41b8-812b-a95a27f72727\") " pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.056907 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/769615ff-2aa5-41b8-812b-a95a27f72727-operator-scripts\") pod \"neutron-c6af-account-create-update-bzzjg\" (UID: \"769615ff-2aa5-41b8-812b-a95a27f72727\") " pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.072377 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll52c\" (UniqueName: \"kubernetes.io/projected/769615ff-2aa5-41b8-812b-a95a27f72727-kube-api-access-ll52c\") pod \"neutron-c6af-account-create-update-bzzjg\" (UID: \"769615ff-2aa5-41b8-812b-a95a27f72727\") " pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.180427 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.191245 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hbltr"] Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.196349 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:19 crc kubenswrapper[4690]: W1203 13:31:19.200114 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda58de3c0_aba8_49ce_a59c_67732420e08f.slice/crio-da4f93c9552396b001e8838eb2c3598411dcd3286fa9432318049bfc673b0e10 WatchSource:0}: Error finding container da4f93c9552396b001e8838eb2c3598411dcd3286fa9432318049bfc673b0e10: Status 404 returned error can't find the container with id da4f93c9552396b001e8838eb2c3598411dcd3286fa9432318049bfc673b0e10 Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.322638 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-gx6kj"] Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.474417 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dab4-account-create-update-445qb"] Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.550752 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4898-account-create-update-66xl7"] Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.636541 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hbltr" event={"ID":"a58de3c0-aba8-49ce-a59c-67732420e08f","Type":"ContainerStarted","Data":"afe960a664ab75914e0e7966e9fce674c927eae62877287a5ba6cb22b82cb458"} Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.636586 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hbltr" event={"ID":"a58de3c0-aba8-49ce-a59c-67732420e08f","Type":"ContainerStarted","Data":"da4f93c9552396b001e8838eb2c3598411dcd3286fa9432318049bfc673b0e10"} Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.662262 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-cg7p2"] Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.672707 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-hbltr" podStartSLOduration=1.6726857050000001 podStartE2EDuration="1.672685705s" podCreationTimestamp="2025-12-03 13:31:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:19.659549199 +0000 UTC m=+1325.640469632" watchObservedRunningTime="2025-12-03 13:31:19.672685705 +0000 UTC m=+1325.653606138" Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.721646 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c6af-account-create-update-bzzjg"] Dec 03 13:31:19 crc kubenswrapper[4690]: I1203 13:31:19.816951 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-59mk8"] Dec 03 13:31:19 crc kubenswrapper[4690]: W1203 13:31:19.985482 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod769615ff_2aa5_41b8_812b_a95a27f72727.slice/crio-c0f3da63efc37c3272e52171e0b0fcaaf073dc75b4621754b147b5305b7799fd WatchSource:0}: Error finding container c0f3da63efc37c3272e52171e0b0fcaaf073dc75b4621754b147b5305b7799fd: Status 404 returned error can't find the container with id c0f3da63efc37c3272e52171e0b0fcaaf073dc75b4621754b147b5305b7799fd Dec 03 13:31:19 crc kubenswrapper[4690]: W1203 13:31:19.986160 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad027298_b2ff_4145_9a45_2dcbf5b5f228.slice/crio-bb535c95988af9acd5c2c8fc5911fe3d7bd9d3245948d46f781fd0ea1d792620 WatchSource:0}: Error finding container bb535c95988af9acd5c2c8fc5911fe3d7bd9d3245948d46f781fd0ea1d792620: Status 404 returned error can't find the container with id bb535c95988af9acd5c2c8fc5911fe3d7bd9d3245948d46f781fd0ea1d792620 Dec 03 13:31:19 crc kubenswrapper[4690]: W1203 13:31:19.987659 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c3e3956_3bd9_4693_9fd1_5040c9f90950.slice/crio-1d94050a0adf474cd7341a905e5f2b5e7b596a85325d5c31a48a7150bb78f688 WatchSource:0}: Error finding container 1d94050a0adf474cd7341a905e5f2b5e7b596a85325d5c31a48a7150bb78f688: Status 404 returned error can't find the container with id 1d94050a0adf474cd7341a905e5f2b5e7b596a85325d5c31a48a7150bb78f688 Dec 03 13:31:19 crc kubenswrapper[4690]: W1203 13:31:19.989186 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7b204d5_9b12_409c_ac51_6a14dc45debe.slice/crio-754c54cd5e79afd99ff5c15d49ff178efe4dbc230a51aa54837db6e00f83e10b WatchSource:0}: Error finding container 754c54cd5e79afd99ff5c15d49ff178efe4dbc230a51aa54837db6e00f83e10b: Status 404 returned error can't find the container with id 754c54cd5e79afd99ff5c15d49ff178efe4dbc230a51aa54837db6e00f83e10b Dec 03 13:31:19 crc kubenswrapper[4690]: W1203 13:31:19.992208 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8441b637_1109_48c4_8bd5_efbcd4a34a56.slice/crio-681f83ae6f31cf8e7945d22d6014922ecc7afa1061c4f8d34c7b05e5ca39cec4 WatchSource:0}: Error finding container 681f83ae6f31cf8e7945d22d6014922ecc7afa1061c4f8d34c7b05e5ca39cec4: Status 404 returned error can't find the container with id 681f83ae6f31cf8e7945d22d6014922ecc7afa1061c4f8d34c7b05e5ca39cec4 Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.001063 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.648244 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dab4-account-create-update-445qb" event={"ID":"4c3e3956-3bd9-4693-9fd1-5040c9f90950","Type":"ContainerStarted","Data":"630a2298b07ad935706b0bfcb75059cf6a14df30bf6a0924854fdcd10094049f"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.648598 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dab4-account-create-update-445qb" event={"ID":"4c3e3956-3bd9-4693-9fd1-5040c9f90950","Type":"ContainerStarted","Data":"1d94050a0adf474cd7341a905e5f2b5e7b596a85325d5c31a48a7150bb78f688"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.650423 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4898-account-create-update-66xl7" event={"ID":"d7b204d5-9b12-409c-ac51-6a14dc45debe","Type":"ContainerStarted","Data":"3a786c0c07d6ddec354cc5ecbe3e7e12dc2108dc64543a9211a0ea72586f7d5b"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.650478 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4898-account-create-update-66xl7" event={"ID":"d7b204d5-9b12-409c-ac51-6a14dc45debe","Type":"ContainerStarted","Data":"754c54cd5e79afd99ff5c15d49ff178efe4dbc230a51aa54837db6e00f83e10b"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.652562 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gx6kj" event={"ID":"8441b637-1109-48c4-8bd5-efbcd4a34a56","Type":"ContainerStarted","Data":"681f83ae6f31cf8e7945d22d6014922ecc7afa1061c4f8d34c7b05e5ca39cec4"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.654676 4690 generic.go:334] "Generic (PLEG): container finished" podID="a58de3c0-aba8-49ce-a59c-67732420e08f" containerID="afe960a664ab75914e0e7966e9fce674c927eae62877287a5ba6cb22b82cb458" exitCode=0 Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.654727 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hbltr" event={"ID":"a58de3c0-aba8-49ce-a59c-67732420e08f","Type":"ContainerDied","Data":"afe960a664ab75914e0e7966e9fce674c927eae62877287a5ba6cb22b82cb458"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.656322 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cg7p2" event={"ID":"153bc447-b9f4-4746-9869-17088a3b71ae","Type":"ContainerStarted","Data":"468fffb5df7b7d90ac81410be0ab3f431f3a1d61d58e777040a814b51319cf54"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.660265 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c6af-account-create-update-bzzjg" event={"ID":"769615ff-2aa5-41b8-812b-a95a27f72727","Type":"ContainerStarted","Data":"b4bf607df358d77e16ffd4970d7128e96cfda86b194a0aeb0d5fea1f0bed72a0"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.660302 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c6af-account-create-update-bzzjg" event={"ID":"769615ff-2aa5-41b8-812b-a95a27f72727","Type":"ContainerStarted","Data":"c0f3da63efc37c3272e52171e0b0fcaaf073dc75b4621754b147b5305b7799fd"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.663598 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-59mk8" event={"ID":"ad027298-b2ff-4145-9a45-2dcbf5b5f228","Type":"ContainerStarted","Data":"c1042ec07c23e9f063277430ca0b7cfd0d67cdcd5665c98a64703096e959dd03"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.663628 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-59mk8" event={"ID":"ad027298-b2ff-4145-9a45-2dcbf5b5f228","Type":"ContainerStarted","Data":"bb535c95988af9acd5c2c8fc5911fe3d7bd9d3245948d46f781fd0ea1d792620"} Dec 03 13:31:20 crc kubenswrapper[4690]: I1203 13:31:20.716369 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-59mk8" podStartSLOduration=2.716338625 podStartE2EDuration="2.716338625s" podCreationTimestamp="2025-12-03 13:31:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:20.700865975 +0000 UTC m=+1326.681786408" watchObservedRunningTime="2025-12-03 13:31:20.716338625 +0000 UTC m=+1326.697259058" Dec 03 13:31:21 crc kubenswrapper[4690]: I1203 13:31:21.674655 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gx6kj" event={"ID":"8441b637-1109-48c4-8bd5-efbcd4a34a56","Type":"ContainerStarted","Data":"cab2422914baf7975e011762ba51c134d42cd25c400e6a0a141d69b1d1845292"} Dec 03 13:31:21 crc kubenswrapper[4690]: I1203 13:31:21.683371 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"34a9215cecf5eb596c4fab2137a1c6187542f36d8590bcec5849fb4a10ceff6d"} Dec 03 13:31:21 crc kubenswrapper[4690]: I1203 13:31:21.698040 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-gx6kj" podStartSLOduration=3.698023642 podStartE2EDuration="3.698023642s" podCreationTimestamp="2025-12-03 13:31:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:21.694630273 +0000 UTC m=+1327.675550706" watchObservedRunningTime="2025-12-03 13:31:21.698023642 +0000 UTC m=+1327.678944075" Dec 03 13:31:21 crc kubenswrapper[4690]: I1203 13:31:21.720673 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c6af-account-create-update-bzzjg" podStartSLOduration=3.720648249 podStartE2EDuration="3.720648249s" podCreationTimestamp="2025-12-03 13:31:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:21.71209365 +0000 UTC m=+1327.693014083" watchObservedRunningTime="2025-12-03 13:31:21.720648249 +0000 UTC m=+1327.701568682" Dec 03 13:31:21 crc kubenswrapper[4690]: I1203 13:31:21.739422 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-dab4-account-create-update-445qb" podStartSLOduration=3.739402776 podStartE2EDuration="3.739402776s" podCreationTimestamp="2025-12-03 13:31:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:21.739321704 +0000 UTC m=+1327.720242137" watchObservedRunningTime="2025-12-03 13:31:21.739402776 +0000 UTC m=+1327.720323209" Dec 03 13:31:21 crc kubenswrapper[4690]: I1203 13:31:21.770805 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-4898-account-create-update-66xl7" podStartSLOduration=3.7707704570000002 podStartE2EDuration="3.770770457s" podCreationTimestamp="2025-12-03 13:31:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:21.763917847 +0000 UTC m=+1327.744838290" watchObservedRunningTime="2025-12-03 13:31:21.770770457 +0000 UTC m=+1327.751690900" Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.148437 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.230082 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a58de3c0-aba8-49ce-a59c-67732420e08f-operator-scripts\") pod \"a58de3c0-aba8-49ce-a59c-67732420e08f\" (UID: \"a58de3c0-aba8-49ce-a59c-67732420e08f\") " Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.230231 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx52v\" (UniqueName: \"kubernetes.io/projected/a58de3c0-aba8-49ce-a59c-67732420e08f-kube-api-access-tx52v\") pod \"a58de3c0-aba8-49ce-a59c-67732420e08f\" (UID: \"a58de3c0-aba8-49ce-a59c-67732420e08f\") " Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.231448 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a58de3c0-aba8-49ce-a59c-67732420e08f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a58de3c0-aba8-49ce-a59c-67732420e08f" (UID: "a58de3c0-aba8-49ce-a59c-67732420e08f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.237507 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a58de3c0-aba8-49ce-a59c-67732420e08f-kube-api-access-tx52v" (OuterVolumeSpecName: "kube-api-access-tx52v") pod "a58de3c0-aba8-49ce-a59c-67732420e08f" (UID: "a58de3c0-aba8-49ce-a59c-67732420e08f"). InnerVolumeSpecName "kube-api-access-tx52v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.332465 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx52v\" (UniqueName: \"kubernetes.io/projected/a58de3c0-aba8-49ce-a59c-67732420e08f-kube-api-access-tx52v\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.332513 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a58de3c0-aba8-49ce-a59c-67732420e08f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.751355 4690 generic.go:334] "Generic (PLEG): container finished" podID="ad027298-b2ff-4145-9a45-2dcbf5b5f228" containerID="c1042ec07c23e9f063277430ca0b7cfd0d67cdcd5665c98a64703096e959dd03" exitCode=0 Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.751425 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-59mk8" event={"ID":"ad027298-b2ff-4145-9a45-2dcbf5b5f228","Type":"ContainerDied","Data":"c1042ec07c23e9f063277430ca0b7cfd0d67cdcd5665c98a64703096e959dd03"} Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.757617 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"cb8b59752f3047a5d6d8ee80512205104acee8659d259721e17a2b63a7c95343"} Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.757656 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"162a96eec7dd86a483abdd8274cb42c8757a987e9f61b4e6eb7bcf7c37346590"} Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.757668 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"baa9338d19817eb33500c52561461c6dc6323b14ea36b484bf4149b802bf167a"} Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.759730 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hbltr" event={"ID":"a58de3c0-aba8-49ce-a59c-67732420e08f","Type":"ContainerDied","Data":"da4f93c9552396b001e8838eb2c3598411dcd3286fa9432318049bfc673b0e10"} Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.759798 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da4f93c9552396b001e8838eb2c3598411dcd3286fa9432318049bfc673b0e10" Dec 03 13:31:22 crc kubenswrapper[4690]: I1203 13:31:22.759746 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hbltr" Dec 03 13:31:23 crc kubenswrapper[4690]: I1203 13:31:23.787205 4690 generic.go:334] "Generic (PLEG): container finished" podID="4c3e3956-3bd9-4693-9fd1-5040c9f90950" containerID="630a2298b07ad935706b0bfcb75059cf6a14df30bf6a0924854fdcd10094049f" exitCode=0 Dec 03 13:31:23 crc kubenswrapper[4690]: I1203 13:31:23.787319 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dab4-account-create-update-445qb" event={"ID":"4c3e3956-3bd9-4693-9fd1-5040c9f90950","Type":"ContainerDied","Data":"630a2298b07ad935706b0bfcb75059cf6a14df30bf6a0924854fdcd10094049f"} Dec 03 13:31:23 crc kubenswrapper[4690]: I1203 13:31:23.799325 4690 generic.go:334] "Generic (PLEG): container finished" podID="8441b637-1109-48c4-8bd5-efbcd4a34a56" containerID="cab2422914baf7975e011762ba51c134d42cd25c400e6a0a141d69b1d1845292" exitCode=0 Dec 03 13:31:23 crc kubenswrapper[4690]: I1203 13:31:23.799509 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gx6kj" event={"ID":"8441b637-1109-48c4-8bd5-efbcd4a34a56","Type":"ContainerDied","Data":"cab2422914baf7975e011762ba51c134d42cd25c400e6a0a141d69b1d1845292"} Dec 03 13:31:24 crc kubenswrapper[4690]: I1203 13:31:24.816718 4690 generic.go:334] "Generic (PLEG): container finished" podID="d7b204d5-9b12-409c-ac51-6a14dc45debe" containerID="3a786c0c07d6ddec354cc5ecbe3e7e12dc2108dc64543a9211a0ea72586f7d5b" exitCode=0 Dec 03 13:31:24 crc kubenswrapper[4690]: I1203 13:31:24.816938 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4898-account-create-update-66xl7" event={"ID":"d7b204d5-9b12-409c-ac51-6a14dc45debe","Type":"ContainerDied","Data":"3a786c0c07d6ddec354cc5ecbe3e7e12dc2108dc64543a9211a0ea72586f7d5b"} Dec 03 13:31:24 crc kubenswrapper[4690]: I1203 13:31:24.824306 4690 generic.go:334] "Generic (PLEG): container finished" podID="769615ff-2aa5-41b8-812b-a95a27f72727" containerID="b4bf607df358d77e16ffd4970d7128e96cfda86b194a0aeb0d5fea1f0bed72a0" exitCode=0 Dec 03 13:31:24 crc kubenswrapper[4690]: I1203 13:31:24.824530 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c6af-account-create-update-bzzjg" event={"ID":"769615ff-2aa5-41b8-812b-a95a27f72727","Type":"ContainerDied","Data":"b4bf607df358d77e16ffd4970d7128e96cfda86b194a0aeb0d5fea1f0bed72a0"} Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.262317 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.269141 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.275268 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.288431 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.298606 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.362401 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp4nk\" (UniqueName: \"kubernetes.io/projected/8441b637-1109-48c4-8bd5-efbcd4a34a56-kube-api-access-sp4nk\") pod \"8441b637-1109-48c4-8bd5-efbcd4a34a56\" (UID: \"8441b637-1109-48c4-8bd5-efbcd4a34a56\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.362455 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/769615ff-2aa5-41b8-812b-a95a27f72727-operator-scripts\") pod \"769615ff-2aa5-41b8-812b-a95a27f72727\" (UID: \"769615ff-2aa5-41b8-812b-a95a27f72727\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.362519 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-864r8\" (UniqueName: \"kubernetes.io/projected/d7b204d5-9b12-409c-ac51-6a14dc45debe-kube-api-access-864r8\") pod \"d7b204d5-9b12-409c-ac51-6a14dc45debe\" (UID: \"d7b204d5-9b12-409c-ac51-6a14dc45debe\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.362629 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfxnf\" (UniqueName: \"kubernetes.io/projected/4c3e3956-3bd9-4693-9fd1-5040c9f90950-kube-api-access-vfxnf\") pod \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\" (UID: \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.362656 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c3e3956-3bd9-4693-9fd1-5040c9f90950-operator-scripts\") pod \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\" (UID: \"4c3e3956-3bd9-4693-9fd1-5040c9f90950\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.362692 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8441b637-1109-48c4-8bd5-efbcd4a34a56-operator-scripts\") pod \"8441b637-1109-48c4-8bd5-efbcd4a34a56\" (UID: \"8441b637-1109-48c4-8bd5-efbcd4a34a56\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.362757 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll52c\" (UniqueName: \"kubernetes.io/projected/769615ff-2aa5-41b8-812b-a95a27f72727-kube-api-access-ll52c\") pod \"769615ff-2aa5-41b8-812b-a95a27f72727\" (UID: \"769615ff-2aa5-41b8-812b-a95a27f72727\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.362822 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b204d5-9b12-409c-ac51-6a14dc45debe-operator-scripts\") pod \"d7b204d5-9b12-409c-ac51-6a14dc45debe\" (UID: \"d7b204d5-9b12-409c-ac51-6a14dc45debe\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.363948 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8441b637-1109-48c4-8bd5-efbcd4a34a56-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8441b637-1109-48c4-8bd5-efbcd4a34a56" (UID: "8441b637-1109-48c4-8bd5-efbcd4a34a56"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.363986 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b204d5-9b12-409c-ac51-6a14dc45debe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7b204d5-9b12-409c-ac51-6a14dc45debe" (UID: "d7b204d5-9b12-409c-ac51-6a14dc45debe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.364759 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/769615ff-2aa5-41b8-812b-a95a27f72727-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "769615ff-2aa5-41b8-812b-a95a27f72727" (UID: "769615ff-2aa5-41b8-812b-a95a27f72727"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.364999 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c3e3956-3bd9-4693-9fd1-5040c9f90950-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c3e3956-3bd9-4693-9fd1-5040c9f90950" (UID: "4c3e3956-3bd9-4693-9fd1-5040c9f90950"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.372183 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/769615ff-2aa5-41b8-812b-a95a27f72727-kube-api-access-ll52c" (OuterVolumeSpecName: "kube-api-access-ll52c") pod "769615ff-2aa5-41b8-812b-a95a27f72727" (UID: "769615ff-2aa5-41b8-812b-a95a27f72727"). InnerVolumeSpecName "kube-api-access-ll52c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.392195 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c3e3956-3bd9-4693-9fd1-5040c9f90950-kube-api-access-vfxnf" (OuterVolumeSpecName: "kube-api-access-vfxnf") pod "4c3e3956-3bd9-4693-9fd1-5040c9f90950" (UID: "4c3e3956-3bd9-4693-9fd1-5040c9f90950"). InnerVolumeSpecName "kube-api-access-vfxnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.393098 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b204d5-9b12-409c-ac51-6a14dc45debe-kube-api-access-864r8" (OuterVolumeSpecName: "kube-api-access-864r8") pod "d7b204d5-9b12-409c-ac51-6a14dc45debe" (UID: "d7b204d5-9b12-409c-ac51-6a14dc45debe"). InnerVolumeSpecName "kube-api-access-864r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.411942 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8441b637-1109-48c4-8bd5-efbcd4a34a56-kube-api-access-sp4nk" (OuterVolumeSpecName: "kube-api-access-sp4nk") pod "8441b637-1109-48c4-8bd5-efbcd4a34a56" (UID: "8441b637-1109-48c4-8bd5-efbcd4a34a56"). InnerVolumeSpecName "kube-api-access-sp4nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.464692 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad027298-b2ff-4145-9a45-2dcbf5b5f228-operator-scripts\") pod \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\" (UID: \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.464795 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d56pj\" (UniqueName: \"kubernetes.io/projected/ad027298-b2ff-4145-9a45-2dcbf5b5f228-kube-api-access-d56pj\") pod \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\" (UID: \"ad027298-b2ff-4145-9a45-2dcbf5b5f228\") " Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465362 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad027298-b2ff-4145-9a45-2dcbf5b5f228-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad027298-b2ff-4145-9a45-2dcbf5b5f228" (UID: "ad027298-b2ff-4145-9a45-2dcbf5b5f228"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465391 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfxnf\" (UniqueName: \"kubernetes.io/projected/4c3e3956-3bd9-4693-9fd1-5040c9f90950-kube-api-access-vfxnf\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465448 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c3e3956-3bd9-4693-9fd1-5040c9f90950-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465461 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8441b637-1109-48c4-8bd5-efbcd4a34a56-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465473 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll52c\" (UniqueName: \"kubernetes.io/projected/769615ff-2aa5-41b8-812b-a95a27f72727-kube-api-access-ll52c\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465484 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b204d5-9b12-409c-ac51-6a14dc45debe-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465496 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp4nk\" (UniqueName: \"kubernetes.io/projected/8441b637-1109-48c4-8bd5-efbcd4a34a56-kube-api-access-sp4nk\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465507 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/769615ff-2aa5-41b8-812b-a95a27f72727-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.465519 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-864r8\" (UniqueName: \"kubernetes.io/projected/d7b204d5-9b12-409c-ac51-6a14dc45debe-kube-api-access-864r8\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.471254 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad027298-b2ff-4145-9a45-2dcbf5b5f228-kube-api-access-d56pj" (OuterVolumeSpecName: "kube-api-access-d56pj") pod "ad027298-b2ff-4145-9a45-2dcbf5b5f228" (UID: "ad027298-b2ff-4145-9a45-2dcbf5b5f228"). InnerVolumeSpecName "kube-api-access-d56pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.570463 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad027298-b2ff-4145-9a45-2dcbf5b5f228-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.570512 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d56pj\" (UniqueName: \"kubernetes.io/projected/ad027298-b2ff-4145-9a45-2dcbf5b5f228-kube-api-access-d56pj\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.873008 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gx6kj" event={"ID":"8441b637-1109-48c4-8bd5-efbcd4a34a56","Type":"ContainerDied","Data":"681f83ae6f31cf8e7945d22d6014922ecc7afa1061c4f8d34c7b05e5ca39cec4"} Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.873395 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="681f83ae6f31cf8e7945d22d6014922ecc7afa1061c4f8d34c7b05e5ca39cec4" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.873053 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gx6kj" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.893283 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cg7p2" event={"ID":"153bc447-b9f4-4746-9869-17088a3b71ae","Type":"ContainerStarted","Data":"5df4f0ea5dcf8aa4c3d6c016fb6a7094f3e4b79be97a3e8e862d4a02fff5e276"} Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.896558 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c6af-account-create-update-bzzjg" event={"ID":"769615ff-2aa5-41b8-812b-a95a27f72727","Type":"ContainerDied","Data":"c0f3da63efc37c3272e52171e0b0fcaaf073dc75b4621754b147b5305b7799fd"} Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.896591 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c6af-account-create-update-bzzjg" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.896613 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0f3da63efc37c3272e52171e0b0fcaaf073dc75b4621754b147b5305b7799fd" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.906228 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-59mk8" event={"ID":"ad027298-b2ff-4145-9a45-2dcbf5b5f228","Type":"ContainerDied","Data":"bb535c95988af9acd5c2c8fc5911fe3d7bd9d3245948d46f781fd0ea1d792620"} Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.906281 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb535c95988af9acd5c2c8fc5911fe3d7bd9d3245948d46f781fd0ea1d792620" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.906371 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-59mk8" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.917652 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-cg7p2" podStartSLOduration=2.455394295 podStartE2EDuration="10.917624751s" podCreationTimestamp="2025-12-03 13:31:18 +0000 UTC" firstStartedPulling="2025-12-03 13:31:20.00075223 +0000 UTC m=+1325.981672683" lastFinishedPulling="2025-12-03 13:31:28.462982706 +0000 UTC m=+1334.443903139" observedRunningTime="2025-12-03 13:31:28.915136683 +0000 UTC m=+1334.896057116" watchObservedRunningTime="2025-12-03 13:31:28.917624751 +0000 UTC m=+1334.898545184" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.918532 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dab4-account-create-update-445qb" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.918574 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dab4-account-create-update-445qb" event={"ID":"4c3e3956-3bd9-4693-9fd1-5040c9f90950","Type":"ContainerDied","Data":"1d94050a0adf474cd7341a905e5f2b5e7b596a85325d5c31a48a7150bb78f688"} Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.918632 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d94050a0adf474cd7341a905e5f2b5e7b596a85325d5c31a48a7150bb78f688" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.934042 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4898-account-create-update-66xl7" event={"ID":"d7b204d5-9b12-409c-ac51-6a14dc45debe","Type":"ContainerDied","Data":"754c54cd5e79afd99ff5c15d49ff178efe4dbc230a51aa54837db6e00f83e10b"} Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.934214 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="754c54cd5e79afd99ff5c15d49ff178efe4dbc230a51aa54837db6e00f83e10b" Dec 03 13:31:28 crc kubenswrapper[4690]: I1203 13:31:28.934319 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4898-account-create-update-66xl7" Dec 03 13:31:29 crc kubenswrapper[4690]: I1203 13:31:29.946362 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"3791f410119864773041047b8ccf97ce1f1b67c4f8af8ac4ccacd8c9052aaeb9"} Dec 03 13:31:29 crc kubenswrapper[4690]: I1203 13:31:29.946409 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"f42dce67ee3d0874f9ecec3b503b4a0512782155acf47ebe0b9057ae57738253"} Dec 03 13:31:30 crc kubenswrapper[4690]: I1203 13:31:30.967561 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"1e82d7961b69e702528bd27c21eb47fdd8fa51787a0aef32bd17894a9a40d959"} Dec 03 13:31:31 crc kubenswrapper[4690]: I1203 13:31:31.985748 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"5cdadeab730de736996030ebba844def2cd751796ad3eda46dd2ed5f8c932bb6"} Dec 03 13:31:34 crc kubenswrapper[4690]: I1203 13:31:34.029041 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"f80a62121ae67785838eda377c06bc95d94135d7ebbebff6d44cd07b56bceb8a"} Dec 03 13:31:34 crc kubenswrapper[4690]: I1203 13:31:34.030444 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fq5zp" event={"ID":"f1a38ea3-0bc0-4533-bbec-de05b19433c7","Type":"ContainerStarted","Data":"da2dee4909c61fe2e7fc186fa657a3df3adf97f4c8a6b05fba3b41a855c10360"} Dec 03 13:31:34 crc kubenswrapper[4690]: I1203 13:31:34.076412 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-fq5zp" podStartSLOduration=3.418848103 podStartE2EDuration="33.076392706s" podCreationTimestamp="2025-12-03 13:31:01 +0000 UTC" firstStartedPulling="2025-12-03 13:31:03.118782668 +0000 UTC m=+1309.099703111" lastFinishedPulling="2025-12-03 13:31:32.776327281 +0000 UTC m=+1338.757247714" observedRunningTime="2025-12-03 13:31:34.058277283 +0000 UTC m=+1340.039197716" watchObservedRunningTime="2025-12-03 13:31:34.076392706 +0000 UTC m=+1340.057313139" Dec 03 13:31:35 crc kubenswrapper[4690]: I1203 13:31:35.048312 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"4f859791445ac5867fab9c9318bdf0def67feda5fe4cf424a16ece2c636a5744"} Dec 03 13:31:36 crc kubenswrapper[4690]: I1203 13:31:36.063222 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"392fac009f3b7ed270394d3f0e97a0ac2994480af247276720457ac4d1b41d19"} Dec 03 13:31:36 crc kubenswrapper[4690]: I1203 13:31:36.063661 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"b87864a3b2bfaa559d0ec5f72403150ca479271a44e4f0937fcdf1757d708669"} Dec 03 13:31:38 crc kubenswrapper[4690]: I1203 13:31:38.099114 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"c69d8df9fc24c39a6566ddbcaf40268214d314a5ff407838bb8cb0595ad2e294"} Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.117347 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"a2197a5ae7663abb83a22ea7b16da985bc7f70aa3bd74498efe016bfab064d16"} Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.117706 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"28748086-659b-453c-bb91-8d9204374da6","Type":"ContainerStarted","Data":"4ac4a91b71cd35798c0e495a66131e374853cea532375d312b0d10fcaf8b46e7"} Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.162823 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=41.498410007 podStartE2EDuration="1m6.162791795s" podCreationTimestamp="2025-12-03 13:30:33 +0000 UTC" firstStartedPulling="2025-12-03 13:31:08.114171325 +0000 UTC m=+1314.095091758" lastFinishedPulling="2025-12-03 13:31:32.778553113 +0000 UTC m=+1338.759473546" observedRunningTime="2025-12-03 13:31:39.153099709 +0000 UTC m=+1345.134020142" watchObservedRunningTime="2025-12-03 13:31:39.162791795 +0000 UTC m=+1345.143712228" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.467276 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fdwgv"] Dec 03 13:31:39 crc kubenswrapper[4690]: E1203 13:31:39.468291 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b204d5-9b12-409c-ac51-6a14dc45debe" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468314 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b204d5-9b12-409c-ac51-6a14dc45debe" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: E1203 13:31:39.468339 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8441b637-1109-48c4-8bd5-efbcd4a34a56" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468347 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8441b637-1109-48c4-8bd5-efbcd4a34a56" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: E1203 13:31:39.468362 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad027298-b2ff-4145-9a45-2dcbf5b5f228" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468369 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad027298-b2ff-4145-9a45-2dcbf5b5f228" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: E1203 13:31:39.468382 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a58de3c0-aba8-49ce-a59c-67732420e08f" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468389 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a58de3c0-aba8-49ce-a59c-67732420e08f" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: E1203 13:31:39.468407 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769615ff-2aa5-41b8-812b-a95a27f72727" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468413 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="769615ff-2aa5-41b8-812b-a95a27f72727" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: E1203 13:31:39.468428 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c3e3956-3bd9-4693-9fd1-5040c9f90950" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468433 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c3e3956-3bd9-4693-9fd1-5040c9f90950" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468637 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a58de3c0-aba8-49ce-a59c-67732420e08f" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468649 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b204d5-9b12-409c-ac51-6a14dc45debe" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468658 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="769615ff-2aa5-41b8-812b-a95a27f72727" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468670 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="8441b637-1109-48c4-8bd5-efbcd4a34a56" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468682 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c3e3956-3bd9-4693-9fd1-5040c9f90950" containerName="mariadb-account-create-update" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.468698 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad027298-b2ff-4145-9a45-2dcbf5b5f228" containerName="mariadb-database-create" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.469965 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.472387 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.492040 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fdwgv"] Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.531343 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-config\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.531604 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.531702 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.532013 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dpsz\" (UniqueName: \"kubernetes.io/projected/5db12289-7ee9-4cfe-aebd-52a40a403569-kube-api-access-9dpsz\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.532165 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-svc\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.532341 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.634141 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.634227 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.634264 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dpsz\" (UniqueName: \"kubernetes.io/projected/5db12289-7ee9-4cfe-aebd-52a40a403569-kube-api-access-9dpsz\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.634306 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-svc\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.634350 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.634380 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-config\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.635663 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.635708 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-config\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.635861 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.635988 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.636325 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-svc\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.663636 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dpsz\" (UniqueName: \"kubernetes.io/projected/5db12289-7ee9-4cfe-aebd-52a40a403569-kube-api-access-9dpsz\") pod \"dnsmasq-dns-764c5664d7-fdwgv\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:39 crc kubenswrapper[4690]: I1203 13:31:39.793365 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:40 crc kubenswrapper[4690]: I1203 13:31:40.533157 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fdwgv"] Dec 03 13:31:41 crc kubenswrapper[4690]: I1203 13:31:41.141801 4690 generic.go:334] "Generic (PLEG): container finished" podID="5db12289-7ee9-4cfe-aebd-52a40a403569" containerID="816d7fb241e9ad68caf954ce96cb33b971f57004e04247ac9a735092d5d2dfad" exitCode=0 Dec 03 13:31:41 crc kubenswrapper[4690]: I1203 13:31:41.141906 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" event={"ID":"5db12289-7ee9-4cfe-aebd-52a40a403569","Type":"ContainerDied","Data":"816d7fb241e9ad68caf954ce96cb33b971f57004e04247ac9a735092d5d2dfad"} Dec 03 13:31:41 crc kubenswrapper[4690]: I1203 13:31:41.142230 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" event={"ID":"5db12289-7ee9-4cfe-aebd-52a40a403569","Type":"ContainerStarted","Data":"d7b7c949d7927a54e7e06855cfb02b417c5d0193bdb4c057da56031fa399708b"} Dec 03 13:31:42 crc kubenswrapper[4690]: I1203 13:31:42.161039 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" event={"ID":"5db12289-7ee9-4cfe-aebd-52a40a403569","Type":"ContainerStarted","Data":"0e3a02ff3c2a32478bc1b77c5aead84dea941240636b315aac3ef90bbdc906f2"} Dec 03 13:31:42 crc kubenswrapper[4690]: I1203 13:31:42.161708 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:42 crc kubenswrapper[4690]: I1203 13:31:42.195756 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" podStartSLOduration=3.19573327 podStartE2EDuration="3.19573327s" podCreationTimestamp="2025-12-03 13:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:42.184091049 +0000 UTC m=+1348.165011492" watchObservedRunningTime="2025-12-03 13:31:42.19573327 +0000 UTC m=+1348.176653713" Dec 03 13:31:44 crc kubenswrapper[4690]: I1203 13:31:44.193582 4690 generic.go:334] "Generic (PLEG): container finished" podID="153bc447-b9f4-4746-9869-17088a3b71ae" containerID="5df4f0ea5dcf8aa4c3d6c016fb6a7094f3e4b79be97a3e8e862d4a02fff5e276" exitCode=0 Dec 03 13:31:44 crc kubenswrapper[4690]: I1203 13:31:44.194260 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cg7p2" event={"ID":"153bc447-b9f4-4746-9869-17088a3b71ae","Type":"ContainerDied","Data":"5df4f0ea5dcf8aa4c3d6c016fb6a7094f3e4b79be97a3e8e862d4a02fff5e276"} Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.579939 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.660542 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdlmn\" (UniqueName: \"kubernetes.io/projected/153bc447-b9f4-4746-9869-17088a3b71ae-kube-api-access-pdlmn\") pod \"153bc447-b9f4-4746-9869-17088a3b71ae\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.661277 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-config-data\") pod \"153bc447-b9f4-4746-9869-17088a3b71ae\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.661329 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-combined-ca-bundle\") pod \"153bc447-b9f4-4746-9869-17088a3b71ae\" (UID: \"153bc447-b9f4-4746-9869-17088a3b71ae\") " Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.678313 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/153bc447-b9f4-4746-9869-17088a3b71ae-kube-api-access-pdlmn" (OuterVolumeSpecName: "kube-api-access-pdlmn") pod "153bc447-b9f4-4746-9869-17088a3b71ae" (UID: "153bc447-b9f4-4746-9869-17088a3b71ae"). InnerVolumeSpecName "kube-api-access-pdlmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.733049 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "153bc447-b9f4-4746-9869-17088a3b71ae" (UID: "153bc447-b9f4-4746-9869-17088a3b71ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.763240 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdlmn\" (UniqueName: \"kubernetes.io/projected/153bc447-b9f4-4746-9869-17088a3b71ae-kube-api-access-pdlmn\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.763821 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.768133 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-config-data" (OuterVolumeSpecName: "config-data") pod "153bc447-b9f4-4746-9869-17088a3b71ae" (UID: "153bc447-b9f4-4746-9869-17088a3b71ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:31:45 crc kubenswrapper[4690]: I1203 13:31:45.866316 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153bc447-b9f4-4746-9869-17088a3b71ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.215678 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cg7p2" event={"ID":"153bc447-b9f4-4746-9869-17088a3b71ae","Type":"ContainerDied","Data":"468fffb5df7b7d90ac81410be0ab3f431f3a1d61d58e777040a814b51319cf54"} Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.215725 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="468fffb5df7b7d90ac81410be0ab3f431f3a1d61d58e777040a814b51319cf54" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.215765 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cg7p2" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.521513 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fdwgv"] Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.521997 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" podUID="5db12289-7ee9-4cfe-aebd-52a40a403569" containerName="dnsmasq-dns" containerID="cri-o://0e3a02ff3c2a32478bc1b77c5aead84dea941240636b315aac3ef90bbdc906f2" gracePeriod=10 Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.523125 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.557499 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wnf2t"] Dec 03 13:31:46 crc kubenswrapper[4690]: E1203 13:31:46.558183 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="153bc447-b9f4-4746-9869-17088a3b71ae" containerName="keystone-db-sync" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.558216 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="153bc447-b9f4-4746-9869-17088a3b71ae" containerName="keystone-db-sync" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.558486 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="153bc447-b9f4-4746-9869-17088a3b71ae" containerName="keystone-db-sync" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.559471 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.564517 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.564758 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.564943 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wpcrf" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.565072 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.565173 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.577846 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wnf2t"] Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.606974 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-mf87t"] Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.609067 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.669957 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-mf87t"] Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686579 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686639 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-fernet-keys\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686656 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-combined-ca-bundle\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686677 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-scripts\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686706 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686734 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8rjb\" (UniqueName: \"kubernetes.io/projected/18b17078-d979-4895-9c23-b0f8d87f858e-kube-api-access-h8rjb\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686754 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-config-data\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686769 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686792 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-config\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686811 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9ljd\" (UniqueName: \"kubernetes.io/projected/2af61167-fd6c-4a12-8e0a-ba02354818b6-kube-api-access-q9ljd\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686889 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-credential-keys\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.686913 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-svc\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.727369 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-2rlzz"] Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.728671 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.734479 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.734678 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-lqrtb" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.734932 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.739511 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-2rlzz"] Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788104 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88d2e02a-af82-49f0-858a-9e9fa3dc4985-etc-machine-id\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788163 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8rjb\" (UniqueName: \"kubernetes.io/projected/18b17078-d979-4895-9c23-b0f8d87f858e-kube-api-access-h8rjb\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788190 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-config-data\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788208 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788234 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-config\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788250 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9ljd\" (UniqueName: \"kubernetes.io/projected/2af61167-fd6c-4a12-8e0a-ba02354818b6-kube-api-access-q9ljd\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788286 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-combined-ca-bundle\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788302 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-config-data\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788341 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk8ps\" (UniqueName: \"kubernetes.io/projected/88d2e02a-af82-49f0-858a-9e9fa3dc4985-kube-api-access-sk8ps\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788381 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-credential-keys\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788397 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-db-sync-config-data\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788420 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-svc\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788454 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788484 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-fernet-keys\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788502 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-combined-ca-bundle\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788520 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-scripts\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788538 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-scripts\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.788569 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.789397 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.790755 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.791504 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-svc\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.791546 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.792194 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-config\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.810200 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-fernet-keys\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.825845 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-config-data\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.826707 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-credential-keys\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.839156 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-scripts\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.839267 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-combined-ca-bundle\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.840531 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8rjb\" (UniqueName: \"kubernetes.io/projected/18b17078-d979-4895-9c23-b0f8d87f858e-kube-api-access-h8rjb\") pod \"keystone-bootstrap-wnf2t\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.870984 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.874148 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.900559 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9ljd\" (UniqueName: \"kubernetes.io/projected/2af61167-fd6c-4a12-8e0a-ba02354818b6-kube-api-access-q9ljd\") pod \"dnsmasq-dns-5959f8865f-mf87t\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.901859 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.902199 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.919338 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-scripts\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.919457 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88d2e02a-af82-49f0-858a-9e9fa3dc4985-etc-machine-id\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.919554 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-combined-ca-bundle\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.919574 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-config-data\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.919637 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk8ps\" (UniqueName: \"kubernetes.io/projected/88d2e02a-af82-49f0-858a-9e9fa3dc4985-kube-api-access-sk8ps\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.919691 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-db-sync-config-data\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.920628 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.921849 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88d2e02a-af82-49f0-858a-9e9fa3dc4985-etc-machine-id\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.937771 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-config-data\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.940656 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.968395 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ndpp9"] Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.973988 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-scripts\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.974485 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-db-sync-config-data\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.978621 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk8ps\" (UniqueName: \"kubernetes.io/projected/88d2e02a-af82-49f0-858a-9e9fa3dc4985-kube-api-access-sk8ps\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.980992 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.994179 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-combined-ca-bundle\") pod \"cinder-db-sync-2rlzz\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.994848 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 13:31:46 crc kubenswrapper[4690]: I1203 13:31:46.995229 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-22rqt" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.032181 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ndpp9"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.033577 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-config-data\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.033693 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.033794 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-log-httpd\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.033822 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.033989 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-scripts\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.034041 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mp8r\" (UniqueName: \"kubernetes.io/projected/942f1c14-f12c-499f-9cd9-934bcdcfee99-kube-api-access-7mp8r\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.034101 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-run-httpd\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.061868 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.072572 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.126432 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2xpmv"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.128650 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.133136 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-q5ggj" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.133579 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.133771 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.151563 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.151665 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4smcb\" (UniqueName: \"kubernetes.io/projected/62230b01-b33e-4213-80c3-ecf6d5f995b1-kube-api-access-4smcb\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.151726 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-log-httpd\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.151747 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.151932 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-scripts\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.151991 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mp8r\" (UniqueName: \"kubernetes.io/projected/942f1c14-f12c-499f-9cd9-934bcdcfee99-kube-api-access-7mp8r\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.152028 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-combined-ca-bundle\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.152068 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-run-httpd\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.152324 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-db-sync-config-data\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.152357 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-config-data\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.153202 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-run-httpd\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.153472 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-log-httpd\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.157809 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-scripts\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.158468 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.159920 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.163904 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-config-data\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.168430 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2xpmv"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.181214 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mp8r\" (UniqueName: \"kubernetes.io/projected/942f1c14-f12c-499f-9cd9-934bcdcfee99-kube-api-access-7mp8r\") pod \"ceilometer-0\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.195940 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-mf87t"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.221377 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-2vcg6"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.222907 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.231542 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2vcg6"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.232266 4690 generic.go:334] "Generic (PLEG): container finished" podID="5db12289-7ee9-4cfe-aebd-52a40a403569" containerID="0e3a02ff3c2a32478bc1b77c5aead84dea941240636b315aac3ef90bbdc906f2" exitCode=0 Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.232298 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" event={"ID":"5db12289-7ee9-4cfe-aebd-52a40a403569","Type":"ContainerDied","Data":"0e3a02ff3c2a32478bc1b77c5aead84dea941240636b315aac3ef90bbdc906f2"} Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.238983 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-gkh4f"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.240735 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.242245 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ckjs2" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.242596 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.242786 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.254316 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4smcb\" (UniqueName: \"kubernetes.io/projected/62230b01-b33e-4213-80c3-ecf6d5f995b1-kube-api-access-4smcb\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.254403 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsfr8\" (UniqueName: \"kubernetes.io/projected/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-kube-api-access-lsfr8\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.254490 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-combined-ca-bundle\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.254516 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-config\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.254551 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-combined-ca-bundle\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.254594 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-db-sync-config-data\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.260992 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-gkh4f"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.264513 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-combined-ca-bundle\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.265028 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-db-sync-config-data\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.291786 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4smcb\" (UniqueName: \"kubernetes.io/projected/62230b01-b33e-4213-80c3-ecf6d5f995b1-kube-api-access-4smcb\") pod \"barbican-db-sync-ndpp9\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357056 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-config\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357124 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357169 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-combined-ca-bundle\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357208 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-config-data\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357237 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-config\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357258 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357298 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6s72\" (UniqueName: \"kubernetes.io/projected/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-kube-api-access-n6s72\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357345 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bf3e45-7939-4b48-923b-d0f28bf5c046-logs\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357403 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsfr8\" (UniqueName: \"kubernetes.io/projected/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-kube-api-access-lsfr8\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357458 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7rzg\" (UniqueName: \"kubernetes.io/projected/29bf3e45-7939-4b48-923b-d0f28bf5c046-kube-api-access-x7rzg\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357488 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-combined-ca-bundle\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357514 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-scripts\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357565 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.357610 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.365286 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-combined-ca-bundle\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.365945 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-config\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.389761 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsfr8\" (UniqueName: \"kubernetes.io/projected/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-kube-api-access-lsfr8\") pod \"neutron-db-sync-2xpmv\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.390278 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.419549 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.464963 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465047 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-config-data\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465089 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-config\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465106 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465157 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6s72\" (UniqueName: \"kubernetes.io/projected/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-kube-api-access-n6s72\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465205 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bf3e45-7939-4b48-923b-d0f28bf5c046-logs\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465273 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7rzg\" (UniqueName: \"kubernetes.io/projected/29bf3e45-7939-4b48-923b-d0f28bf5c046-kube-api-access-x7rzg\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465296 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-combined-ca-bundle\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465315 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-scripts\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465339 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.465364 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.466147 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.469490 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.477595 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-config\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.478142 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.479207 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.480737 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bf3e45-7939-4b48-923b-d0f28bf5c046-logs\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.484214 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7rzg\" (UniqueName: \"kubernetes.io/projected/29bf3e45-7939-4b48-923b-d0f28bf5c046-kube-api-access-x7rzg\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.485123 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-scripts\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.485351 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-config-data\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.486399 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-combined-ca-bundle\") pod \"placement-db-sync-2vcg6\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.486582 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.503385 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6s72\" (UniqueName: \"kubernetes.io/projected/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-kube-api-access-n6s72\") pod \"dnsmasq-dns-58dd9ff6bc-gkh4f\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.592376 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vcg6" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.618700 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.692983 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wnf2t"] Dec 03 13:31:47 crc kubenswrapper[4690]: I1203 13:31:47.980446 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-2rlzz"] Dec 03 13:31:48 crc kubenswrapper[4690]: W1203 13:31:48.279262 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18b17078_d979_4895_9c23_b0f8d87f858e.slice/crio-16f0bd6dae90e50788eaf9d5d3e77ed1de5a4be61bc1d0e93bff211a37691434 WatchSource:0}: Error finding container 16f0bd6dae90e50788eaf9d5d3e77ed1de5a4be61bc1d0e93bff211a37691434: Status 404 returned error can't find the container with id 16f0bd6dae90e50788eaf9d5d3e77ed1de5a4be61bc1d0e93bff211a37691434 Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.313342 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-mf87t"] Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.507376 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.595736 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dpsz\" (UniqueName: \"kubernetes.io/projected/5db12289-7ee9-4cfe-aebd-52a40a403569-kube-api-access-9dpsz\") pod \"5db12289-7ee9-4cfe-aebd-52a40a403569\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.595890 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-nb\") pod \"5db12289-7ee9-4cfe-aebd-52a40a403569\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.596845 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-svc\") pod \"5db12289-7ee9-4cfe-aebd-52a40a403569\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.597104 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-sb\") pod \"5db12289-7ee9-4cfe-aebd-52a40a403569\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.597222 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-config\") pod \"5db12289-7ee9-4cfe-aebd-52a40a403569\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.597287 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-swift-storage-0\") pod \"5db12289-7ee9-4cfe-aebd-52a40a403569\" (UID: \"5db12289-7ee9-4cfe-aebd-52a40a403569\") " Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.606980 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5db12289-7ee9-4cfe-aebd-52a40a403569-kube-api-access-9dpsz" (OuterVolumeSpecName: "kube-api-access-9dpsz") pod "5db12289-7ee9-4cfe-aebd-52a40a403569" (UID: "5db12289-7ee9-4cfe-aebd-52a40a403569"). InnerVolumeSpecName "kube-api-access-9dpsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.702018 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dpsz\" (UniqueName: \"kubernetes.io/projected/5db12289-7ee9-4cfe-aebd-52a40a403569-kube-api-access-9dpsz\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.704468 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5db12289-7ee9-4cfe-aebd-52a40a403569" (UID: "5db12289-7ee9-4cfe-aebd-52a40a403569"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.730750 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5db12289-7ee9-4cfe-aebd-52a40a403569" (UID: "5db12289-7ee9-4cfe-aebd-52a40a403569"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.774339 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-config" (OuterVolumeSpecName: "config") pod "5db12289-7ee9-4cfe-aebd-52a40a403569" (UID: "5db12289-7ee9-4cfe-aebd-52a40a403569"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.780973 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5db12289-7ee9-4cfe-aebd-52a40a403569" (UID: "5db12289-7ee9-4cfe-aebd-52a40a403569"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.789036 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5db12289-7ee9-4cfe-aebd-52a40a403569" (UID: "5db12289-7ee9-4cfe-aebd-52a40a403569"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.804285 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.804316 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.804327 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.804338 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.804350 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db12289-7ee9-4cfe-aebd-52a40a403569-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:48 crc kubenswrapper[4690]: I1203 13:31:48.881222 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ndpp9"] Dec 03 13:31:48 crc kubenswrapper[4690]: W1203 13:31:48.887532 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62230b01_b33e_4213_80c3_ecf6d5f995b1.slice/crio-9d23ee71098c333ecf04b12d677c1953379bd96e598dc886319183111d92c0ae WatchSource:0}: Error finding container 9d23ee71098c333ecf04b12d677c1953379bd96e598dc886319183111d92c0ae: Status 404 returned error can't find the container with id 9d23ee71098c333ecf04b12d677c1953379bd96e598dc886319183111d92c0ae Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.120542 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2vcg6"] Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.136341 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2xpmv"] Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.166218 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.263030 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2xpmv" event={"ID":"3ca321dd-bd9d-4039-aa80-eb386d63dcb4","Type":"ContainerStarted","Data":"8fdeff4385a0edc212a40024bd94720d185c7facc4fd42829603ba46d6d7b9fb"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.267453 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" event={"ID":"5db12289-7ee9-4cfe-aebd-52a40a403569","Type":"ContainerDied","Data":"d7b7c949d7927a54e7e06855cfb02b417c5d0193bdb4c057da56031fa399708b"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.267497 4690 scope.go:117] "RemoveContainer" containerID="0e3a02ff3c2a32478bc1b77c5aead84dea941240636b315aac3ef90bbdc906f2" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.267684 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-fdwgv" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.271445 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"942f1c14-f12c-499f-9cd9-934bcdcfee99","Type":"ContainerStarted","Data":"77a82477e445c6a37bb46a1f9d80462479f3139cb4d8fa063fd3adc0cbe79976"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.274023 4690 generic.go:334] "Generic (PLEG): container finished" podID="2af61167-fd6c-4a12-8e0a-ba02354818b6" containerID="d7a713a64683af44da4eb886c4fa8e90e713c0c5d0455881f39d2254e80867ce" exitCode=0 Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.274086 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-mf87t" event={"ID":"2af61167-fd6c-4a12-8e0a-ba02354818b6","Type":"ContainerDied","Data":"d7a713a64683af44da4eb886c4fa8e90e713c0c5d0455881f39d2254e80867ce"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.274112 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-mf87t" event={"ID":"2af61167-fd6c-4a12-8e0a-ba02354818b6","Type":"ContainerStarted","Data":"3742bb3f7df035548a1a7e462af4aef0133acb008314ced381a5714835c8e0c1"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.280154 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wnf2t" event={"ID":"18b17078-d979-4895-9c23-b0f8d87f858e","Type":"ContainerStarted","Data":"72e74369053341505d34978936a6f7ddf7df8ec8c983484863ea6f3ec0a362ea"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.280200 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wnf2t" event={"ID":"18b17078-d979-4895-9c23-b0f8d87f858e","Type":"ContainerStarted","Data":"16f0bd6dae90e50788eaf9d5d3e77ed1de5a4be61bc1d0e93bff211a37691434"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.282344 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ndpp9" event={"ID":"62230b01-b33e-4213-80c3-ecf6d5f995b1","Type":"ContainerStarted","Data":"9d23ee71098c333ecf04b12d677c1953379bd96e598dc886319183111d92c0ae"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.283369 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vcg6" event={"ID":"29bf3e45-7939-4b48-923b-d0f28bf5c046","Type":"ContainerStarted","Data":"3de4a911250866c5c93cc5ee2b044f75158f7189df676aa06fcabd9dbeb9626c"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.284230 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2rlzz" event={"ID":"88d2e02a-af82-49f0-858a-9e9fa3dc4985","Type":"ContainerStarted","Data":"a82ed63ec10f55dcee76d3065d231f9b07e90270c8ab3d45613f47efc9e874f1"} Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.339924 4690 scope.go:117] "RemoveContainer" containerID="816d7fb241e9ad68caf954ce96cb33b971f57004e04247ac9a735092d5d2dfad" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.392242 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wnf2t" podStartSLOduration=3.39221844 podStartE2EDuration="3.39221844s" podCreationTimestamp="2025-12-03 13:31:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:49.327956302 +0000 UTC m=+1355.308876755" watchObservedRunningTime="2025-12-03 13:31:49.39221844 +0000 UTC m=+1355.373138873" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.435957 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fdwgv"] Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.449324 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-gkh4f"] Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.462813 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fdwgv"] Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.662485 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.772034 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.831415 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-nb\") pod \"2af61167-fd6c-4a12-8e0a-ba02354818b6\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.832377 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-config\") pod \"2af61167-fd6c-4a12-8e0a-ba02354818b6\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.832484 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-svc\") pod \"2af61167-fd6c-4a12-8e0a-ba02354818b6\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.832651 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-sb\") pod \"2af61167-fd6c-4a12-8e0a-ba02354818b6\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.832744 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-swift-storage-0\") pod \"2af61167-fd6c-4a12-8e0a-ba02354818b6\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.832917 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9ljd\" (UniqueName: \"kubernetes.io/projected/2af61167-fd6c-4a12-8e0a-ba02354818b6-kube-api-access-q9ljd\") pod \"2af61167-fd6c-4a12-8e0a-ba02354818b6\" (UID: \"2af61167-fd6c-4a12-8e0a-ba02354818b6\") " Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.841552 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2af61167-fd6c-4a12-8e0a-ba02354818b6-kube-api-access-q9ljd" (OuterVolumeSpecName: "kube-api-access-q9ljd") pod "2af61167-fd6c-4a12-8e0a-ba02354818b6" (UID: "2af61167-fd6c-4a12-8e0a-ba02354818b6"). InnerVolumeSpecName "kube-api-access-q9ljd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.893112 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2af61167-fd6c-4a12-8e0a-ba02354818b6" (UID: "2af61167-fd6c-4a12-8e0a-ba02354818b6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.903392 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-config" (OuterVolumeSpecName: "config") pod "2af61167-fd6c-4a12-8e0a-ba02354818b6" (UID: "2af61167-fd6c-4a12-8e0a-ba02354818b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.904696 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2af61167-fd6c-4a12-8e0a-ba02354818b6" (UID: "2af61167-fd6c-4a12-8e0a-ba02354818b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.937419 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9ljd\" (UniqueName: \"kubernetes.io/projected/2af61167-fd6c-4a12-8e0a-ba02354818b6-kube-api-access-q9ljd\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.937466 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.937477 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.937488 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.966042 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2af61167-fd6c-4a12-8e0a-ba02354818b6" (UID: "2af61167-fd6c-4a12-8e0a-ba02354818b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:49 crc kubenswrapper[4690]: I1203 13:31:49.985911 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2af61167-fd6c-4a12-8e0a-ba02354818b6" (UID: "2af61167-fd6c-4a12-8e0a-ba02354818b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.039641 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.039680 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2af61167-fd6c-4a12-8e0a-ba02354818b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.294123 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2xpmv" event={"ID":"3ca321dd-bd9d-4039-aa80-eb386d63dcb4","Type":"ContainerStarted","Data":"eb7088271fec7391af00bc7bb58cc71b4f9de4efd677b58db0288059fc39f036"} Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.299172 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" event={"ID":"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b","Type":"ContainerStarted","Data":"ff1781670859a363e94b9b97a46cffda2a5bc9a2910db781dda9118fe7dd5431"} Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.299229 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" event={"ID":"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b","Type":"ContainerStarted","Data":"f3572dc8becd0c9bab703cbcb104210f2c34dd5198b9756160cc14c7a3c984d6"} Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.303263 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-mf87t" event={"ID":"2af61167-fd6c-4a12-8e0a-ba02354818b6","Type":"ContainerDied","Data":"3742bb3f7df035548a1a7e462af4aef0133acb008314ced381a5714835c8e0c1"} Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.303328 4690 scope.go:117] "RemoveContainer" containerID="d7a713a64683af44da4eb886c4fa8e90e713c0c5d0455881f39d2254e80867ce" Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.303436 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-mf87t" Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.317647 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2xpmv" podStartSLOduration=4.317626645 podStartE2EDuration="4.317626645s" podCreationTimestamp="2025-12-03 13:31:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:50.31655632 +0000 UTC m=+1356.297476753" watchObservedRunningTime="2025-12-03 13:31:50.317626645 +0000 UTC m=+1356.298547078" Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.329393 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5db12289-7ee9-4cfe-aebd-52a40a403569" path="/var/lib/kubelet/pods/5db12289-7ee9-4cfe-aebd-52a40a403569/volumes" Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.379398 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-mf87t"] Dec 03 13:31:50 crc kubenswrapper[4690]: I1203 13:31:50.398202 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-mf87t"] Dec 03 13:31:51 crc kubenswrapper[4690]: I1203 13:31:51.324835 4690 generic.go:334] "Generic (PLEG): container finished" podID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerID="ff1781670859a363e94b9b97a46cffda2a5bc9a2910db781dda9118fe7dd5431" exitCode=0 Dec 03 13:31:51 crc kubenswrapper[4690]: I1203 13:31:51.325004 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" event={"ID":"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b","Type":"ContainerDied","Data":"ff1781670859a363e94b9b97a46cffda2a5bc9a2910db781dda9118fe7dd5431"} Dec 03 13:31:52 crc kubenswrapper[4690]: I1203 13:31:52.329601 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2af61167-fd6c-4a12-8e0a-ba02354818b6" path="/var/lib/kubelet/pods/2af61167-fd6c-4a12-8e0a-ba02354818b6/volumes" Dec 03 13:31:52 crc kubenswrapper[4690]: I1203 13:31:52.337845 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" event={"ID":"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b","Type":"ContainerStarted","Data":"c5d320031c5314650cf936d492d9012062039ef119d0a1c04f3a620dc99aa3dd"} Dec 03 13:31:52 crc kubenswrapper[4690]: I1203 13:31:52.338013 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:52 crc kubenswrapper[4690]: I1203 13:31:52.385810 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" podStartSLOduration=5.385794849 podStartE2EDuration="5.385794849s" podCreationTimestamp="2025-12-03 13:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:31:52.364109864 +0000 UTC m=+1358.345030297" watchObservedRunningTime="2025-12-03 13:31:52.385794849 +0000 UTC m=+1358.366715282" Dec 03 13:31:57 crc kubenswrapper[4690]: I1203 13:31:57.621072 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:31:57 crc kubenswrapper[4690]: I1203 13:31:57.695261 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qtkhc"] Dec 03 13:31:57 crc kubenswrapper[4690]: I1203 13:31:57.695502 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-qtkhc" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="dnsmasq-dns" containerID="cri-o://54660f3c082858a9860457fc52b1df5e7817a7957a36fb967cb370ed62948347" gracePeriod=10 Dec 03 13:31:58 crc kubenswrapper[4690]: I1203 13:31:58.409359 4690 generic.go:334] "Generic (PLEG): container finished" podID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerID="54660f3c082858a9860457fc52b1df5e7817a7957a36fb967cb370ed62948347" exitCode=0 Dec 03 13:31:58 crc kubenswrapper[4690]: I1203 13:31:58.409421 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qtkhc" event={"ID":"aed37082-28cb-49e4-9d87-e8246df4b55c","Type":"ContainerDied","Data":"54660f3c082858a9860457fc52b1df5e7817a7957a36fb967cb370ed62948347"} Dec 03 13:31:58 crc kubenswrapper[4690]: I1203 13:31:58.949978 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qtkhc" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Dec 03 13:31:59 crc kubenswrapper[4690]: I1203 13:31:59.420262 4690 generic.go:334] "Generic (PLEG): container finished" podID="f1a38ea3-0bc0-4533-bbec-de05b19433c7" containerID="da2dee4909c61fe2e7fc186fa657a3df3adf97f4c8a6b05fba3b41a855c10360" exitCode=0 Dec 03 13:31:59 crc kubenswrapper[4690]: I1203 13:31:59.420368 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fq5zp" event={"ID":"f1a38ea3-0bc0-4533-bbec-de05b19433c7","Type":"ContainerDied","Data":"da2dee4909c61fe2e7fc186fa657a3df3adf97f4c8a6b05fba3b41a855c10360"} Dec 03 13:31:59 crc kubenswrapper[4690]: I1203 13:31:59.421693 4690 generic.go:334] "Generic (PLEG): container finished" podID="18b17078-d979-4895-9c23-b0f8d87f858e" containerID="72e74369053341505d34978936a6f7ddf7df8ec8c983484863ea6f3ec0a362ea" exitCode=0 Dec 03 13:31:59 crc kubenswrapper[4690]: I1203 13:31:59.421715 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wnf2t" event={"ID":"18b17078-d979-4895-9c23-b0f8d87f858e","Type":"ContainerDied","Data":"72e74369053341505d34978936a6f7ddf7df8ec8c983484863ea6f3ec0a362ea"} Dec 03 13:32:08 crc kubenswrapper[4690]: I1203 13:32:08.949073 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qtkhc" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: i/o timeout" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.469541 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.478163 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fq5zp" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.486533 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.540457 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fq5zp" event={"ID":"f1a38ea3-0bc0-4533-bbec-de05b19433c7","Type":"ContainerDied","Data":"7c53356369419b0fce73a9f5bddd4dcd219016815079ed4075e31d588d2fea33"} Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.540517 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c53356369419b0fce73a9f5bddd4dcd219016815079ed4075e31d588d2fea33" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.540577 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fq5zp" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.551623 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wnf2t" event={"ID":"18b17078-d979-4895-9c23-b0f8d87f858e","Type":"ContainerDied","Data":"16f0bd6dae90e50788eaf9d5d3e77ed1de5a4be61bc1d0e93bff211a37691434"} Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.551675 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16f0bd6dae90e50788eaf9d5d3e77ed1de5a4be61bc1d0e93bff211a37691434" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.551646 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wnf2t" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.563327 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qtkhc" event={"ID":"aed37082-28cb-49e4-9d87-e8246df4b55c","Type":"ContainerDied","Data":"dc1bf79a1e689f1d33e78d95502bd9485e7a3b6081cfa3986a0465ea1f232c9d"} Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.563387 4690 scope.go:117] "RemoveContainer" containerID="54660f3c082858a9860457fc52b1df5e7817a7957a36fb967cb370ed62948347" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.563545 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qtkhc" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630328 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgs72\" (UniqueName: \"kubernetes.io/projected/aed37082-28cb-49e4-9d87-e8246df4b55c-kube-api-access-jgs72\") pod \"aed37082-28cb-49e4-9d87-e8246df4b55c\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630430 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-dns-svc\") pod \"aed37082-28cb-49e4-9d87-e8246df4b55c\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630452 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-scripts\") pod \"18b17078-d979-4895-9c23-b0f8d87f858e\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630491 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-combined-ca-bundle\") pod \"18b17078-d979-4895-9c23-b0f8d87f858e\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630580 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh7vd\" (UniqueName: \"kubernetes.io/projected/f1a38ea3-0bc0-4533-bbec-de05b19433c7-kube-api-access-dh7vd\") pod \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630621 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-config-data\") pod \"18b17078-d979-4895-9c23-b0f8d87f858e\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630638 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-nb\") pod \"aed37082-28cb-49e4-9d87-e8246df4b55c\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630657 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8rjb\" (UniqueName: \"kubernetes.io/projected/18b17078-d979-4895-9c23-b0f8d87f858e-kube-api-access-h8rjb\") pod \"18b17078-d979-4895-9c23-b0f8d87f858e\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630693 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-config-data\") pod \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630729 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-config\") pod \"aed37082-28cb-49e4-9d87-e8246df4b55c\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630746 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-sb\") pod \"aed37082-28cb-49e4-9d87-e8246df4b55c\" (UID: \"aed37082-28cb-49e4-9d87-e8246df4b55c\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630770 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-combined-ca-bundle\") pod \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630790 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-credential-keys\") pod \"18b17078-d979-4895-9c23-b0f8d87f858e\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630840 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-db-sync-config-data\") pod \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\" (UID: \"f1a38ea3-0bc0-4533-bbec-de05b19433c7\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.630957 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-fernet-keys\") pod \"18b17078-d979-4895-9c23-b0f8d87f858e\" (UID: \"18b17078-d979-4895-9c23-b0f8d87f858e\") " Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.639042 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-scripts" (OuterVolumeSpecName: "scripts") pod "18b17078-d979-4895-9c23-b0f8d87f858e" (UID: "18b17078-d979-4895-9c23-b0f8d87f858e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.640444 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed37082-28cb-49e4-9d87-e8246df4b55c-kube-api-access-jgs72" (OuterVolumeSpecName: "kube-api-access-jgs72") pod "aed37082-28cb-49e4-9d87-e8246df4b55c" (UID: "aed37082-28cb-49e4-9d87-e8246df4b55c"). InnerVolumeSpecName "kube-api-access-jgs72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.640510 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1a38ea3-0bc0-4533-bbec-de05b19433c7-kube-api-access-dh7vd" (OuterVolumeSpecName: "kube-api-access-dh7vd") pod "f1a38ea3-0bc0-4533-bbec-de05b19433c7" (UID: "f1a38ea3-0bc0-4533-bbec-de05b19433c7"). InnerVolumeSpecName "kube-api-access-dh7vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.640745 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18b17078-d979-4895-9c23-b0f8d87f858e-kube-api-access-h8rjb" (OuterVolumeSpecName: "kube-api-access-h8rjb") pod "18b17078-d979-4895-9c23-b0f8d87f858e" (UID: "18b17078-d979-4895-9c23-b0f8d87f858e"). InnerVolumeSpecName "kube-api-access-h8rjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.641588 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "18b17078-d979-4895-9c23-b0f8d87f858e" (UID: "18b17078-d979-4895-9c23-b0f8d87f858e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.642410 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "18b17078-d979-4895-9c23-b0f8d87f858e" (UID: "18b17078-d979-4895-9c23-b0f8d87f858e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.645427 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f1a38ea3-0bc0-4533-bbec-de05b19433c7" (UID: "f1a38ea3-0bc0-4533-bbec-de05b19433c7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.665513 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1a38ea3-0bc0-4533-bbec-de05b19433c7" (UID: "f1a38ea3-0bc0-4533-bbec-de05b19433c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.687536 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-config-data" (OuterVolumeSpecName: "config-data") pod "18b17078-d979-4895-9c23-b0f8d87f858e" (UID: "18b17078-d979-4895-9c23-b0f8d87f858e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.689008 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18b17078-d979-4895-9c23-b0f8d87f858e" (UID: "18b17078-d979-4895-9c23-b0f8d87f858e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.689414 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aed37082-28cb-49e4-9d87-e8246df4b55c" (UID: "aed37082-28cb-49e4-9d87-e8246df4b55c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.695578 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-config" (OuterVolumeSpecName: "config") pod "aed37082-28cb-49e4-9d87-e8246df4b55c" (UID: "aed37082-28cb-49e4-9d87-e8246df4b55c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.695689 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aed37082-28cb-49e4-9d87-e8246df4b55c" (UID: "aed37082-28cb-49e4-9d87-e8246df4b55c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.701831 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-config-data" (OuterVolumeSpecName: "config-data") pod "f1a38ea3-0bc0-4533-bbec-de05b19433c7" (UID: "f1a38ea3-0bc0-4533-bbec-de05b19433c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.706461 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aed37082-28cb-49e4-9d87-e8246df4b55c" (UID: "aed37082-28cb-49e4-9d87-e8246df4b55c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.733927 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.733958 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.733970 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.733983 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh7vd\" (UniqueName: \"kubernetes.io/projected/f1a38ea3-0bc0-4533-bbec-de05b19433c7-kube-api-access-dh7vd\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.733993 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734003 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734011 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8rjb\" (UniqueName: \"kubernetes.io/projected/18b17078-d979-4895-9c23-b0f8d87f858e-kube-api-access-h8rjb\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734018 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734026 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734033 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed37082-28cb-49e4-9d87-e8246df4b55c-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734041 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734048 4690 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734056 4690 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1a38ea3-0bc0-4533-bbec-de05b19433c7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734063 4690 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18b17078-d979-4895-9c23-b0f8d87f858e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.734079 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgs72\" (UniqueName: \"kubernetes.io/projected/aed37082-28cb-49e4-9d87-e8246df4b55c-kube-api-access-jgs72\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.902229 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qtkhc"] Dec 03 13:32:11 crc kubenswrapper[4690]: I1203 13:32:11.910461 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qtkhc"] Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.182299 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.182527 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n556hb8h88h575h56hf6h59h5b9h9fh8bh544h65dhfbh86h5bh574h655h54ch6dh5d4h5chbh5bch689h5fdh9bh56fh5f7h5d4h594h665hddq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mp8r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(942f1c14-f12c-499f-9cd9-934bcdcfee99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.333147 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" path="/var/lib/kubelet/pods/aed37082-28cb-49e4-9d87-e8246df4b55c/volumes" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.685515 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wnf2t"] Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.695047 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wnf2t"] Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.781889 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9srzs"] Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.782322 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a38ea3-0bc0-4533-bbec-de05b19433c7" containerName="glance-db-sync" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782345 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a38ea3-0bc0-4533-bbec-de05b19433c7" containerName="glance-db-sync" Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.782364 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db12289-7ee9-4cfe-aebd-52a40a403569" containerName="dnsmasq-dns" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782375 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db12289-7ee9-4cfe-aebd-52a40a403569" containerName="dnsmasq-dns" Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.782385 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18b17078-d979-4895-9c23-b0f8d87f858e" containerName="keystone-bootstrap" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782393 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="18b17078-d979-4895-9c23-b0f8d87f858e" containerName="keystone-bootstrap" Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.782415 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="dnsmasq-dns" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782423 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="dnsmasq-dns" Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.782432 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af61167-fd6c-4a12-8e0a-ba02354818b6" containerName="init" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782439 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af61167-fd6c-4a12-8e0a-ba02354818b6" containerName="init" Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.782452 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="init" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782458 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="init" Dec 03 13:32:12 crc kubenswrapper[4690]: E1203 13:32:12.782478 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db12289-7ee9-4cfe-aebd-52a40a403569" containerName="init" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782486 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db12289-7ee9-4cfe-aebd-52a40a403569" containerName="init" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782650 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1a38ea3-0bc0-4533-bbec-de05b19433c7" containerName="glance-db-sync" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782665 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="dnsmasq-dns" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782682 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af61167-fd6c-4a12-8e0a-ba02354818b6" containerName="init" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782692 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="5db12289-7ee9-4cfe-aebd-52a40a403569" containerName="dnsmasq-dns" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.782701 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="18b17078-d979-4895-9c23-b0f8d87f858e" containerName="keystone-bootstrap" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.783326 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.797651 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.797892 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.797973 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wpcrf" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.798101 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.798174 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.833291 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9srzs"] Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.946011 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-j89sv"] Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.958655 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.960080 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-j89sv"] Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.960614 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-scripts\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.960683 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-combined-ca-bundle\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.960757 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-fernet-keys\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.960816 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th4hn\" (UniqueName: \"kubernetes.io/projected/e754a7fd-8dd6-4ac4-b197-9c26c946d540-kube-api-access-th4hn\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.960845 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-config-data\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:12 crc kubenswrapper[4690]: I1203 13:32:12.960920 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-credential-keys\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063193 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slnfw\" (UniqueName: \"kubernetes.io/projected/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-kube-api-access-slnfw\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063323 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-config\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063374 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-scripts\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063396 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-combined-ca-bundle\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063454 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-fernet-keys\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063489 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063508 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063536 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th4hn\" (UniqueName: \"kubernetes.io/projected/e754a7fd-8dd6-4ac4-b197-9c26c946d540-kube-api-access-th4hn\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063556 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-config-data\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063587 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-credential-keys\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063623 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.063661 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.069373 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-scripts\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.074019 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-fernet-keys\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.074202 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-credential-keys\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.075357 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-config-data\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.090936 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-combined-ca-bundle\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.091442 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th4hn\" (UniqueName: \"kubernetes.io/projected/e754a7fd-8dd6-4ac4-b197-9c26c946d540-kube-api-access-th4hn\") pod \"keystone-bootstrap-9srzs\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.136265 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.165106 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.165152 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.165206 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.165229 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.165274 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slnfw\" (UniqueName: \"kubernetes.io/projected/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-kube-api-access-slnfw\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.165312 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-config\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.166105 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.166277 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.166732 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.166973 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-config\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.167275 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.186853 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slnfw\" (UniqueName: \"kubernetes.io/projected/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-kube-api-access-slnfw\") pod \"dnsmasq-dns-785d8bcb8c-j89sv\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.287105 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.659681 4690 scope.go:117] "RemoveContainer" containerID="cc10549ac276b05176be491d8718bf387f431c1346f8f41c57001a377a09f8a0" Dec 03 13:32:13 crc kubenswrapper[4690]: E1203 13:32:13.854077 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 13:32:13 crc kubenswrapper[4690]: E1203 13:32:13.854607 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sk8ps,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-2rlzz_openstack(88d2e02a-af82-49f0-858a-9e9fa3dc4985): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:32:13 crc kubenswrapper[4690]: E1203 13:32:13.856460 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-2rlzz" podUID="88d2e02a-af82-49f0-858a-9e9fa3dc4985" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.873627 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.875137 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.882193 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.882558 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.887287 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ld4lm" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.900929 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.950447 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qtkhc" podUID="aed37082-28cb-49e4-9d87-e8246df4b55c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: i/o timeout" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.985071 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.985137 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.985264 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-config-data\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.985298 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvbnn\" (UniqueName: \"kubernetes.io/projected/16e41a4b-9d7f-4733-826e-e93b82748cc1-kube-api-access-cvbnn\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.985327 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-logs\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.985388 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:13 crc kubenswrapper[4690]: I1203 13:32:13.985452 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-scripts\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.093776 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.094941 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-config-data\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.095005 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvbnn\" (UniqueName: \"kubernetes.io/projected/16e41a4b-9d7f-4733-826e-e93b82748cc1-kube-api-access-cvbnn\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.095039 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-logs\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.095096 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.095159 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-scripts\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.095721 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-logs\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.095901 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.095204 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.096283 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.096727 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.098848 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.104050 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-config-data\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.104228 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.104392 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-scripts\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.110206 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.116015 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.124425 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvbnn\" (UniqueName: \"kubernetes.io/projected/16e41a4b-9d7f-4733-826e-e93b82748cc1-kube-api-access-cvbnn\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.160689 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.197946 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mk7z\" (UniqueName: \"kubernetes.io/projected/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-kube-api-access-2mk7z\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.198024 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-logs\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.198055 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.198084 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.198155 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.198237 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.198312 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.219664 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.236797 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9srzs"] Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.272692 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.300515 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.300613 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mk7z\" (UniqueName: \"kubernetes.io/projected/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-kube-api-access-2mk7z\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.300658 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-logs\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.300680 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.300706 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.300752 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.300823 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.301513 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-logs\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.301546 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.302077 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.323431 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.349438 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mk7z\" (UniqueName: \"kubernetes.io/projected/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-kube-api-access-2mk7z\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.353514 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.365055 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.368658 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.404744 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18b17078-d979-4895-9c23-b0f8d87f858e" path="/var/lib/kubelet/pods/18b17078-d979-4895-9c23-b0f8d87f858e/volumes" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.411881 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-j89sv"] Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.422405 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.538463 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.610309 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" event={"ID":"c8fd9d7d-4064-48c5-9a0a-cabf1de25647","Type":"ContainerStarted","Data":"5e945d883d0dd47bc0c38178bb170b27117a0bac07d6efe0aeca799cc21d4884"} Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.614093 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ndpp9" event={"ID":"62230b01-b33e-4213-80c3-ecf6d5f995b1","Type":"ContainerStarted","Data":"be3917a1f502330f72af2766300a0420d1fbdd574129993d9051d6ac51c7b363"} Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.618168 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9srzs" event={"ID":"e754a7fd-8dd6-4ac4-b197-9c26c946d540","Type":"ContainerStarted","Data":"d16672f2582c5966174e61a505e383d3b887b1e12db1ff3a2fcbbd6160fbddfb"} Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.627792 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vcg6" event={"ID":"29bf3e45-7939-4b48-923b-d0f28bf5c046","Type":"ContainerStarted","Data":"84d532dd714b5dbf5bc6ee00db6886ddf3ebe64dfec1465c5d61f5ac86193c30"} Dec 03 13:32:14 crc kubenswrapper[4690]: E1203 13:32:14.630194 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-2rlzz" podUID="88d2e02a-af82-49f0-858a-9e9fa3dc4985" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.638158 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ndpp9" podStartSLOduration=3.872918495 podStartE2EDuration="28.638132435s" podCreationTimestamp="2025-12-03 13:31:46 +0000 UTC" firstStartedPulling="2025-12-03 13:31:48.89400333 +0000 UTC m=+1354.874923773" lastFinishedPulling="2025-12-03 13:32:13.65921728 +0000 UTC m=+1379.640137713" observedRunningTime="2025-12-03 13:32:14.634147922 +0000 UTC m=+1380.615068355" watchObservedRunningTime="2025-12-03 13:32:14.638132435 +0000 UTC m=+1380.619052868" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.662679 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9srzs" podStartSLOduration=2.6626609 podStartE2EDuration="2.6626609s" podCreationTimestamp="2025-12-03 13:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:14.658137703 +0000 UTC m=+1380.639058146" watchObservedRunningTime="2025-12-03 13:32:14.6626609 +0000 UTC m=+1380.643581333" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.683664 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-2vcg6" podStartSLOduration=4.224154897 podStartE2EDuration="28.683639663s" podCreationTimestamp="2025-12-03 13:31:46 +0000 UTC" firstStartedPulling="2025-12-03 13:31:49.199886868 +0000 UTC m=+1355.180807301" lastFinishedPulling="2025-12-03 13:32:13.659371634 +0000 UTC m=+1379.640292067" observedRunningTime="2025-12-03 13:32:14.672557306 +0000 UTC m=+1380.653477759" watchObservedRunningTime="2025-12-03 13:32:14.683639663 +0000 UTC m=+1380.664560096" Dec 03 13:32:14 crc kubenswrapper[4690]: I1203 13:32:14.864718 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:14 crc kubenswrapper[4690]: W1203 13:32:14.871668 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16e41a4b_9d7f_4733_826e_e93b82748cc1.slice/crio-b89fb5db1769d580f196b3f92b7128940f00abc5a34515d1d22772dedb1eb4ca WatchSource:0}: Error finding container b89fb5db1769d580f196b3f92b7128940f00abc5a34515d1d22772dedb1eb4ca: Status 404 returned error can't find the container with id b89fb5db1769d580f196b3f92b7128940f00abc5a34515d1d22772dedb1eb4ca Dec 03 13:32:15 crc kubenswrapper[4690]: I1203 13:32:15.283732 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:15 crc kubenswrapper[4690]: I1203 13:32:15.641688 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9srzs" event={"ID":"e754a7fd-8dd6-4ac4-b197-9c26c946d540","Type":"ContainerStarted","Data":"e686b923d12a143f1dff83d5a8f41d5c3566f5144b2f0b912ca84f5d26ac33ef"} Dec 03 13:32:15 crc kubenswrapper[4690]: I1203 13:32:15.648965 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerID="1786ab286f9279ade15677762d98310b86563a2576124c5135af41965bd238b7" exitCode=0 Dec 03 13:32:15 crc kubenswrapper[4690]: I1203 13:32:15.649077 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" event={"ID":"c8fd9d7d-4064-48c5-9a0a-cabf1de25647","Type":"ContainerDied","Data":"1786ab286f9279ade15677762d98310b86563a2576124c5135af41965bd238b7"} Dec 03 13:32:15 crc kubenswrapper[4690]: I1203 13:32:15.656623 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16e41a4b-9d7f-4733-826e-e93b82748cc1","Type":"ContainerStarted","Data":"b89fb5db1769d580f196b3f92b7128940f00abc5a34515d1d22772dedb1eb4ca"} Dec 03 13:32:15 crc kubenswrapper[4690]: I1203 13:32:15.931573 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:15 crc kubenswrapper[4690]: I1203 13:32:15.999119 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:16 crc kubenswrapper[4690]: I1203 13:32:16.697856 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16e41a4b-9d7f-4733-826e-e93b82748cc1","Type":"ContainerStarted","Data":"3066111e54d3636a0f7a8aa8efb1820da1cf19a0279bf5bce987d68db132aa8a"} Dec 03 13:32:16 crc kubenswrapper[4690]: I1203 13:32:16.700593 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"097f4fa0-dda4-4794-b2ca-1ff25654bb7c","Type":"ContainerStarted","Data":"f22bb7235061e7285bd3386d012ab20ee8a9ddc37b2193677afc7c2a25578563"} Dec 03 13:32:17 crc kubenswrapper[4690]: I1203 13:32:17.711883 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" event={"ID":"c8fd9d7d-4064-48c5-9a0a-cabf1de25647","Type":"ContainerStarted","Data":"c0f7708e37c77380e77a4e58092d11da4ff74a489c99344b91225a9976f62aa1"} Dec 03 13:32:17 crc kubenswrapper[4690]: I1203 13:32:17.712331 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:17 crc kubenswrapper[4690]: I1203 13:32:17.715173 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16e41a4b-9d7f-4733-826e-e93b82748cc1","Type":"ContainerStarted","Data":"d257ec74c7862d87c152fedda9c0dca3339e63b47bda761ae73d66d268263277"} Dec 03 13:32:17 crc kubenswrapper[4690]: I1203 13:32:17.716632 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"097f4fa0-dda4-4794-b2ca-1ff25654bb7c","Type":"ContainerStarted","Data":"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214"} Dec 03 13:32:17 crc kubenswrapper[4690]: I1203 13:32:17.731799 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" podStartSLOduration=5.7317793 podStartE2EDuration="5.7317793s" podCreationTimestamp="2025-12-03 13:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:17.730957529 +0000 UTC m=+1383.711877962" watchObservedRunningTime="2025-12-03 13:32:17.7317793 +0000 UTC m=+1383.712699733" Dec 03 13:32:18 crc kubenswrapper[4690]: I1203 13:32:18.733283 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"097f4fa0-dda4-4794-b2ca-1ff25654bb7c","Type":"ContainerStarted","Data":"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b"} Dec 03 13:32:18 crc kubenswrapper[4690]: I1203 13:32:18.733639 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerName="glance-log" containerID="cri-o://6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214" gracePeriod=30 Dec 03 13:32:18 crc kubenswrapper[4690]: I1203 13:32:18.733849 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerName="glance-httpd" containerID="cri-o://0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b" gracePeriod=30 Dec 03 13:32:18 crc kubenswrapper[4690]: I1203 13:32:18.740818 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerName="glance-log" containerID="cri-o://3066111e54d3636a0f7a8aa8efb1820da1cf19a0279bf5bce987d68db132aa8a" gracePeriod=30 Dec 03 13:32:18 crc kubenswrapper[4690]: I1203 13:32:18.740925 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"942f1c14-f12c-499f-9cd9-934bcdcfee99","Type":"ContainerStarted","Data":"fe584c98cbe2ae53557fa5355768068e0600db371b4937efe754653e8ac46240"} Dec 03 13:32:18 crc kubenswrapper[4690]: I1203 13:32:18.741101 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerName="glance-httpd" containerID="cri-o://d257ec74c7862d87c152fedda9c0dca3339e63b47bda761ae73d66d268263277" gracePeriod=30 Dec 03 13:32:18 crc kubenswrapper[4690]: I1203 13:32:18.764592 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.764569089 podStartE2EDuration="5.764569089s" podCreationTimestamp="2025-12-03 13:32:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:18.756785368 +0000 UTC m=+1384.737705801" watchObservedRunningTime="2025-12-03 13:32:18.764569089 +0000 UTC m=+1384.745489532" Dec 03 13:32:18 crc kubenswrapper[4690]: I1203 13:32:18.795175 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.795151181 podStartE2EDuration="6.795151181s" podCreationTimestamp="2025-12-03 13:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:18.794451953 +0000 UTC m=+1384.775372406" watchObservedRunningTime="2025-12-03 13:32:18.795151181 +0000 UTC m=+1384.776071614" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.501573 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640127 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mk7z\" (UniqueName: \"kubernetes.io/projected/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-kube-api-access-2mk7z\") pod \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640335 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-scripts\") pod \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640504 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-httpd-run\") pod \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640556 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-logs\") pod \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640602 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640711 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-combined-ca-bundle\") pod \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640805 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-config-data\") pod \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\" (UID: \"097f4fa0-dda4-4794-b2ca-1ff25654bb7c\") " Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640828 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-logs" (OuterVolumeSpecName: "logs") pod "097f4fa0-dda4-4794-b2ca-1ff25654bb7c" (UID: "097f4fa0-dda4-4794-b2ca-1ff25654bb7c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.640908 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "097f4fa0-dda4-4794-b2ca-1ff25654bb7c" (UID: "097f4fa0-dda4-4794-b2ca-1ff25654bb7c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.641579 4690 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.641615 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.646904 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "097f4fa0-dda4-4794-b2ca-1ff25654bb7c" (UID: "097f4fa0-dda4-4794-b2ca-1ff25654bb7c"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.647335 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-scripts" (OuterVolumeSpecName: "scripts") pod "097f4fa0-dda4-4794-b2ca-1ff25654bb7c" (UID: "097f4fa0-dda4-4794-b2ca-1ff25654bb7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.647403 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-kube-api-access-2mk7z" (OuterVolumeSpecName: "kube-api-access-2mk7z") pod "097f4fa0-dda4-4794-b2ca-1ff25654bb7c" (UID: "097f4fa0-dda4-4794-b2ca-1ff25654bb7c"). InnerVolumeSpecName "kube-api-access-2mk7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.678611 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "097f4fa0-dda4-4794-b2ca-1ff25654bb7c" (UID: "097f4fa0-dda4-4794-b2ca-1ff25654bb7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.708305 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-config-data" (OuterVolumeSpecName: "config-data") pod "097f4fa0-dda4-4794-b2ca-1ff25654bb7c" (UID: "097f4fa0-dda4-4794-b2ca-1ff25654bb7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.742901 4690 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.742941 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.742956 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.742968 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mk7z\" (UniqueName: \"kubernetes.io/projected/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-kube-api-access-2mk7z\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.742976 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097f4fa0-dda4-4794-b2ca-1ff25654bb7c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.759223 4690 generic.go:334] "Generic (PLEG): container finished" podID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerID="d257ec74c7862d87c152fedda9c0dca3339e63b47bda761ae73d66d268263277" exitCode=0 Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.759265 4690 generic.go:334] "Generic (PLEG): container finished" podID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerID="3066111e54d3636a0f7a8aa8efb1820da1cf19a0279bf5bce987d68db132aa8a" exitCode=143 Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.759314 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16e41a4b-9d7f-4733-826e-e93b82748cc1","Type":"ContainerDied","Data":"d257ec74c7862d87c152fedda9c0dca3339e63b47bda761ae73d66d268263277"} Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.759374 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16e41a4b-9d7f-4733-826e-e93b82748cc1","Type":"ContainerDied","Data":"3066111e54d3636a0f7a8aa8efb1820da1cf19a0279bf5bce987d68db132aa8a"} Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.762694 4690 generic.go:334] "Generic (PLEG): container finished" podID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerID="0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b" exitCode=143 Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.762740 4690 generic.go:334] "Generic (PLEG): container finished" podID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerID="6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214" exitCode=143 Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.762809 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.762828 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"097f4fa0-dda4-4794-b2ca-1ff25654bb7c","Type":"ContainerDied","Data":"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b"} Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.762903 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"097f4fa0-dda4-4794-b2ca-1ff25654bb7c","Type":"ContainerDied","Data":"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214"} Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.762915 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"097f4fa0-dda4-4794-b2ca-1ff25654bb7c","Type":"ContainerDied","Data":"f22bb7235061e7285bd3386d012ab20ee8a9ddc37b2193677afc7c2a25578563"} Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.762933 4690 scope.go:117] "RemoveContainer" containerID="0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.771747 4690 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.783478 4690 generic.go:334] "Generic (PLEG): container finished" podID="e754a7fd-8dd6-4ac4-b197-9c26c946d540" containerID="e686b923d12a143f1dff83d5a8f41d5c3566f5144b2f0b912ca84f5d26ac33ef" exitCode=0 Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.783554 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9srzs" event={"ID":"e754a7fd-8dd6-4ac4-b197-9c26c946d540","Type":"ContainerDied","Data":"e686b923d12a143f1dff83d5a8f41d5c3566f5144b2f0b912ca84f5d26ac33ef"} Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.814377 4690 scope.go:117] "RemoveContainer" containerID="6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.839120 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.846354 4690 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.856999 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.861961 4690 scope.go:117] "RemoveContainer" containerID="0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b" Dec 03 13:32:19 crc kubenswrapper[4690]: E1203 13:32:19.862946 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b\": container with ID starting with 0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b not found: ID does not exist" containerID="0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.863082 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b"} err="failed to get container status \"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b\": rpc error: code = NotFound desc = could not find container \"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b\": container with ID starting with 0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b not found: ID does not exist" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.863121 4690 scope.go:117] "RemoveContainer" containerID="6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214" Dec 03 13:32:19 crc kubenswrapper[4690]: E1203 13:32:19.864219 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214\": container with ID starting with 6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214 not found: ID does not exist" containerID="6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.864267 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214"} err="failed to get container status \"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214\": rpc error: code = NotFound desc = could not find container \"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214\": container with ID starting with 6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214 not found: ID does not exist" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.864305 4690 scope.go:117] "RemoveContainer" containerID="0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.864878 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b"} err="failed to get container status \"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b\": rpc error: code = NotFound desc = could not find container \"0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b\": container with ID starting with 0c457a8da12b2050eb936f502232692acb1b01c9de557acb1800fff95f40415b not found: ID does not exist" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.864901 4690 scope.go:117] "RemoveContainer" containerID="6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.865315 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214"} err="failed to get container status \"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214\": rpc error: code = NotFound desc = could not find container \"6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214\": container with ID starting with 6c4a24493974db36641ba20bb339bbad8ce9a6e8ce199dce329b6393b7e5b214 not found: ID does not exist" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.879929 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:19 crc kubenswrapper[4690]: E1203 13:32:19.880580 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerName="glance-httpd" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.880721 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerName="glance-httpd" Dec 03 13:32:19 crc kubenswrapper[4690]: E1203 13:32:19.880779 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerName="glance-log" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.880789 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerName="glance-log" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.881166 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerName="glance-log" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.881192 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" containerName="glance-httpd" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.882772 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.887397 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.887687 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.895103 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.948574 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.948649 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-logs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.948770 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-config-data\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.948797 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.948819 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-scripts\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.948861 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf7bs\" (UniqueName: \"kubernetes.io/projected/105e22f6-36c9-4ab3-8e64-9f056a266704-kube-api-access-vf7bs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.948899 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:19 crc kubenswrapper[4690]: I1203 13:32:19.948953 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.050490 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.050908 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.050951 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-logs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.051096 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-config-data\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.051127 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.051149 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-scripts\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.051193 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf7bs\" (UniqueName: \"kubernetes.io/projected/105e22f6-36c9-4ab3-8e64-9f056a266704-kube-api-access-vf7bs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.051211 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.051238 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.052231 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.052258 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-logs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.056589 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.056621 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.076004 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-config-data\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.078545 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf7bs\" (UniqueName: \"kubernetes.io/projected/105e22f6-36c9-4ab3-8e64-9f056a266704-kube-api-access-vf7bs\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.082423 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-scripts\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.120437 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.226044 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.331193 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="097f4fa0-dda4-4794-b2ca-1ff25654bb7c" path="/var/lib/kubelet/pods/097f4fa0-dda4-4794-b2ca-1ff25654bb7c/volumes" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.378293 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.460556 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvbnn\" (UniqueName: \"kubernetes.io/projected/16e41a4b-9d7f-4733-826e-e93b82748cc1-kube-api-access-cvbnn\") pod \"16e41a4b-9d7f-4733-826e-e93b82748cc1\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.460959 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-scripts\") pod \"16e41a4b-9d7f-4733-826e-e93b82748cc1\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.461082 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-combined-ca-bundle\") pod \"16e41a4b-9d7f-4733-826e-e93b82748cc1\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.461113 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"16e41a4b-9d7f-4733-826e-e93b82748cc1\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.461168 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-logs\") pod \"16e41a4b-9d7f-4733-826e-e93b82748cc1\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.461324 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-config-data\") pod \"16e41a4b-9d7f-4733-826e-e93b82748cc1\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.461357 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-httpd-run\") pod \"16e41a4b-9d7f-4733-826e-e93b82748cc1\" (UID: \"16e41a4b-9d7f-4733-826e-e93b82748cc1\") " Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.461789 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-logs" (OuterVolumeSpecName: "logs") pod "16e41a4b-9d7f-4733-826e-e93b82748cc1" (UID: "16e41a4b-9d7f-4733-826e-e93b82748cc1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.462107 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "16e41a4b-9d7f-4733-826e-e93b82748cc1" (UID: "16e41a4b-9d7f-4733-826e-e93b82748cc1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.462626 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.462643 4690 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16e41a4b-9d7f-4733-826e-e93b82748cc1-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.470005 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "16e41a4b-9d7f-4733-826e-e93b82748cc1" (UID: "16e41a4b-9d7f-4733-826e-e93b82748cc1"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.470018 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-scripts" (OuterVolumeSpecName: "scripts") pod "16e41a4b-9d7f-4733-826e-e93b82748cc1" (UID: "16e41a4b-9d7f-4733-826e-e93b82748cc1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.470089 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16e41a4b-9d7f-4733-826e-e93b82748cc1-kube-api-access-cvbnn" (OuterVolumeSpecName: "kube-api-access-cvbnn") pod "16e41a4b-9d7f-4733-826e-e93b82748cc1" (UID: "16e41a4b-9d7f-4733-826e-e93b82748cc1"). InnerVolumeSpecName "kube-api-access-cvbnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.503122 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16e41a4b-9d7f-4733-826e-e93b82748cc1" (UID: "16e41a4b-9d7f-4733-826e-e93b82748cc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.530389 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-config-data" (OuterVolumeSpecName: "config-data") pod "16e41a4b-9d7f-4733-826e-e93b82748cc1" (UID: "16e41a4b-9d7f-4733-826e-e93b82748cc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.565559 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.565596 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvbnn\" (UniqueName: \"kubernetes.io/projected/16e41a4b-9d7f-4733-826e-e93b82748cc1-kube-api-access-cvbnn\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.565614 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.565626 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16e41a4b-9d7f-4733-826e-e93b82748cc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.565665 4690 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.587029 4690 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.668006 4690 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.804142 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.804655 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16e41a4b-9d7f-4733-826e-e93b82748cc1","Type":"ContainerDied","Data":"b89fb5db1769d580f196b3f92b7128940f00abc5a34515d1d22772dedb1eb4ca"} Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.804706 4690 scope.go:117] "RemoveContainer" containerID="d257ec74c7862d87c152fedda9c0dca3339e63b47bda761ae73d66d268263277" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.804813 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.846545 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.857655 4690 scope.go:117] "RemoveContainer" containerID="3066111e54d3636a0f7a8aa8efb1820da1cf19a0279bf5bce987d68db132aa8a" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.867778 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.881180 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:20 crc kubenswrapper[4690]: E1203 13:32:20.881655 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerName="glance-httpd" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.881673 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerName="glance-httpd" Dec 03 13:32:20 crc kubenswrapper[4690]: E1203 13:32:20.881689 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerName="glance-log" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.881695 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerName="glance-log" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.881959 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerName="glance-httpd" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.881998 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" containerName="glance-log" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.883375 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.894639 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.896291 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.902020 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.975399 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.975939 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-logs\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.976010 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.976067 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-config-data\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.976095 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.976167 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.976198 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-scripts\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:20 crc kubenswrapper[4690]: I1203 13:32:20.976228 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwmd6\" (UniqueName: \"kubernetes.io/projected/604fdece-2b3e-4786-995d-68d784e5d263-kube-api-access-pwmd6\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.097833 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-config-data\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.098069 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.098145 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.098178 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-scripts\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.098218 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwmd6\" (UniqueName: \"kubernetes.io/projected/604fdece-2b3e-4786-995d-68d784e5d263-kube-api-access-pwmd6\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.098263 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.098332 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-logs\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.098408 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.098925 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.100519 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.105800 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-logs\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.108017 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-config-data\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.112611 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-scripts\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.113468 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.130114 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwmd6\" (UniqueName: \"kubernetes.io/projected/604fdece-2b3e-4786-995d-68d784e5d263-kube-api-access-pwmd6\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.147857 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.169829 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.224257 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.330916 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.405026 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th4hn\" (UniqueName: \"kubernetes.io/projected/e754a7fd-8dd6-4ac4-b197-9c26c946d540-kube-api-access-th4hn\") pod \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.405589 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-combined-ca-bundle\") pod \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.405973 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-fernet-keys\") pod \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.406089 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-scripts\") pod \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.406291 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-config-data\") pod \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.406361 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-credential-keys\") pod \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\" (UID: \"e754a7fd-8dd6-4ac4-b197-9c26c946d540\") " Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.417786 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-scripts" (OuterVolumeSpecName: "scripts") pod "e754a7fd-8dd6-4ac4-b197-9c26c946d540" (UID: "e754a7fd-8dd6-4ac4-b197-9c26c946d540"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.420024 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e754a7fd-8dd6-4ac4-b197-9c26c946d540-kube-api-access-th4hn" (OuterVolumeSpecName: "kube-api-access-th4hn") pod "e754a7fd-8dd6-4ac4-b197-9c26c946d540" (UID: "e754a7fd-8dd6-4ac4-b197-9c26c946d540"). InnerVolumeSpecName "kube-api-access-th4hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.421139 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e754a7fd-8dd6-4ac4-b197-9c26c946d540" (UID: "e754a7fd-8dd6-4ac4-b197-9c26c946d540"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.433537 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e754a7fd-8dd6-4ac4-b197-9c26c946d540" (UID: "e754a7fd-8dd6-4ac4-b197-9c26c946d540"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.457612 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-config-data" (OuterVolumeSpecName: "config-data") pod "e754a7fd-8dd6-4ac4-b197-9c26c946d540" (UID: "e754a7fd-8dd6-4ac4-b197-9c26c946d540"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.463847 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e754a7fd-8dd6-4ac4-b197-9c26c946d540" (UID: "e754a7fd-8dd6-4ac4-b197-9c26c946d540"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.513685 4690 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.513725 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.513734 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.513744 4690 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.513758 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th4hn\" (UniqueName: \"kubernetes.io/projected/e754a7fd-8dd6-4ac4-b197-9c26c946d540-kube-api-access-th4hn\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.513768 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e754a7fd-8dd6-4ac4-b197-9c26c946d540-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.829487 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9srzs" event={"ID":"e754a7fd-8dd6-4ac4-b197-9c26c946d540","Type":"ContainerDied","Data":"d16672f2582c5966174e61a505e383d3b887b1e12db1ff3a2fcbbd6160fbddfb"} Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.834891 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d16672f2582c5966174e61a505e383d3b887b1e12db1ff3a2fcbbd6160fbddfb" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.835094 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9srzs" Dec 03 13:32:21 crc kubenswrapper[4690]: W1203 13:32:21.838959 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604fdece_2b3e_4786_995d_68d784e5d263.slice/crio-9d0d9e980146105b0e63b7a4cb5e155396d85ab2d3eaa573643f6d48db602fb4 WatchSource:0}: Error finding container 9d0d9e980146105b0e63b7a4cb5e155396d85ab2d3eaa573643f6d48db602fb4: Status 404 returned error can't find the container with id 9d0d9e980146105b0e63b7a4cb5e155396d85ab2d3eaa573643f6d48db602fb4 Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.839094 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"105e22f6-36c9-4ab3-8e64-9f056a266704","Type":"ContainerStarted","Data":"a137ccf12ffa8a24112e602c0772b929c1098bb353882e08f88a421e0461a011"} Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.859352 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.993560 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-788df6b698-7gkrp"] Dec 03 13:32:21 crc kubenswrapper[4690]: E1203 13:32:21.994103 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e754a7fd-8dd6-4ac4-b197-9c26c946d540" containerName="keystone-bootstrap" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.994122 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e754a7fd-8dd6-4ac4-b197-9c26c946d540" containerName="keystone-bootstrap" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.994328 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e754a7fd-8dd6-4ac4-b197-9c26c946d540" containerName="keystone-bootstrap" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.995085 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.997161 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.998485 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.998688 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.998884 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.999070 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 13:32:21 crc kubenswrapper[4690]: I1203 13:32:21.999230 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wpcrf" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.004764 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-788df6b698-7gkrp"] Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.127176 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-credential-keys\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.127226 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-combined-ca-bundle\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.127264 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-fernet-keys\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.127302 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-internal-tls-certs\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.127607 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-config-data\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.127649 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckczj\" (UniqueName: \"kubernetes.io/projected/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-kube-api-access-ckczj\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.127703 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-scripts\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.127751 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-public-tls-certs\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.229309 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-scripts\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.229390 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-public-tls-certs\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.229423 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-credential-keys\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.229442 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-combined-ca-bundle\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.229467 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-fernet-keys\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.229494 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-internal-tls-certs\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.229572 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-config-data\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.229601 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckczj\" (UniqueName: \"kubernetes.io/projected/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-kube-api-access-ckczj\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.235648 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-credential-keys\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.235672 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-public-tls-certs\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.236278 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-scripts\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.236945 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-fernet-keys\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.238273 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-combined-ca-bundle\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.238561 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-config-data\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.239183 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-internal-tls-certs\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.246755 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckczj\" (UniqueName: \"kubernetes.io/projected/cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4-kube-api-access-ckczj\") pod \"keystone-788df6b698-7gkrp\" (UID: \"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4\") " pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.320829 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.333440 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16e41a4b-9d7f-4733-826e-e93b82748cc1" path="/var/lib/kubelet/pods/16e41a4b-9d7f-4733-826e-e93b82748cc1/volumes" Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.831987 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-788df6b698-7gkrp"] Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.861399 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"604fdece-2b3e-4786-995d-68d784e5d263","Type":"ContainerStarted","Data":"9d0d9e980146105b0e63b7a4cb5e155396d85ab2d3eaa573643f6d48db602fb4"} Dec 03 13:32:22 crc kubenswrapper[4690]: I1203 13:32:22.863929 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-788df6b698-7gkrp" event={"ID":"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4","Type":"ContainerStarted","Data":"8fda9c37f67de15592f1123acfb46a31c0b3dc00940aeab9069b4466a4520d46"} Dec 03 13:32:23 crc kubenswrapper[4690]: I1203 13:32:23.289936 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:32:23 crc kubenswrapper[4690]: I1203 13:32:23.361123 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-gkh4f"] Dec 03 13:32:23 crc kubenswrapper[4690]: I1203 13:32:23.361664 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="dnsmasq-dns" containerID="cri-o://c5d320031c5314650cf936d492d9012062039ef119d0a1c04f3a620dc99aa3dd" gracePeriod=10 Dec 03 13:32:26 crc kubenswrapper[4690]: I1203 13:32:26.907323 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-788df6b698-7gkrp" event={"ID":"cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4","Type":"ContainerStarted","Data":"8239d1db5e297ea00400c6df650de7d9074de89f82beb5f743d957c5b7940859"} Dec 03 13:32:26 crc kubenswrapper[4690]: I1203 13:32:26.909488 4690 generic.go:334] "Generic (PLEG): container finished" podID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerID="c5d320031c5314650cf936d492d9012062039ef119d0a1c04f3a620dc99aa3dd" exitCode=0 Dec 03 13:32:26 crc kubenswrapper[4690]: I1203 13:32:26.909554 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" event={"ID":"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b","Type":"ContainerDied","Data":"c5d320031c5314650cf936d492d9012062039ef119d0a1c04f3a620dc99aa3dd"} Dec 03 13:32:26 crc kubenswrapper[4690]: I1203 13:32:26.911041 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"604fdece-2b3e-4786-995d-68d784e5d263","Type":"ContainerStarted","Data":"6146c38d3aacfc70340716008a2714e4faa89e7d667db18a56c938df48dbf6d0"} Dec 03 13:32:26 crc kubenswrapper[4690]: I1203 13:32:26.912409 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"105e22f6-36c9-4ab3-8e64-9f056a266704","Type":"ContainerStarted","Data":"49017db7e96374547b2aab24e42d9d98a9647f40c8f34a5ef74506765384840c"} Dec 03 13:32:27 crc kubenswrapper[4690]: I1203 13:32:27.620201 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: connect: connection refused" Dec 03 13:32:27 crc kubenswrapper[4690]: I1203 13:32:27.924078 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:27 crc kubenswrapper[4690]: I1203 13:32:27.956807 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-788df6b698-7gkrp" podStartSLOduration=6.956787928 podStartE2EDuration="6.956787928s" podCreationTimestamp="2025-12-03 13:32:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:27.945489875 +0000 UTC m=+1393.926410328" watchObservedRunningTime="2025-12-03 13:32:27.956787928 +0000 UTC m=+1393.937708361" Dec 03 13:32:28 crc kubenswrapper[4690]: I1203 13:32:28.934766 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"105e22f6-36c9-4ab3-8e64-9f056a266704","Type":"ContainerStarted","Data":"50fb8514b26b86d905c80d505bfd3f7977b619a006d3c4bd64746853de8bcb53"} Dec 03 13:32:28 crc kubenswrapper[4690]: I1203 13:32:28.938431 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"604fdece-2b3e-4786-995d-68d784e5d263","Type":"ContainerStarted","Data":"d030d61740116392c7f29accd15b1953269757d2f02a2a14d263e9faf60b1675"} Dec 03 13:32:36 crc kubenswrapper[4690]: I1203 13:32:36.059908 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=16.059860108 podStartE2EDuration="16.059860108s" podCreationTimestamp="2025-12-03 13:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:36.046818911 +0000 UTC m=+1402.027739354" watchObservedRunningTime="2025-12-03 13:32:36.059860108 +0000 UTC m=+1402.040780541" Dec 03 13:32:36 crc kubenswrapper[4690]: I1203 13:32:36.079335 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.079304662 podStartE2EDuration="17.079304662s" podCreationTimestamp="2025-12-03 13:32:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:36.070963176 +0000 UTC m=+1402.051883609" watchObservedRunningTime="2025-12-03 13:32:36.079304662 +0000 UTC m=+1402.060225095" Dec 03 13:32:37 crc kubenswrapper[4690]: I1203 13:32:37.619756 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: i/o timeout" Dec 03 13:32:38 crc kubenswrapper[4690]: E1203 13:32:38.574818 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Dec 03 13:32:38 crc kubenswrapper[4690]: E1203 13:32:38.575002 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mp8r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(942f1c14-f12c-499f-9cd9-934bcdcfee99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.641895 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.776799 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-config\") pod \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.777123 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6s72\" (UniqueName: \"kubernetes.io/projected/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-kube-api-access-n6s72\") pod \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.777303 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-svc\") pod \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.777396 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-swift-storage-0\") pod \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.777508 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-sb\") pod \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.777622 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-nb\") pod \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\" (UID: \"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b\") " Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.782706 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-kube-api-access-n6s72" (OuterVolumeSpecName: "kube-api-access-n6s72") pod "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" (UID: "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b"). InnerVolumeSpecName "kube-api-access-n6s72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.874553 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" (UID: "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.874610 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" (UID: "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.876295 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" (UID: "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.880216 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6s72\" (UniqueName: \"kubernetes.io/projected/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-kube-api-access-n6s72\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.880252 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.880263 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.880272 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.883130 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" (UID: "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.890176 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-config" (OuterVolumeSpecName: "config") pod "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" (UID: "cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.983780 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:38 crc kubenswrapper[4690]: I1203 13:32:38.983844 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:39 crc kubenswrapper[4690]: I1203 13:32:39.060502 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" event={"ID":"cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b","Type":"ContainerDied","Data":"f3572dc8becd0c9bab703cbcb104210f2c34dd5198b9756160cc14c7a3c984d6"} Dec 03 13:32:39 crc kubenswrapper[4690]: I1203 13:32:39.060579 4690 scope.go:117] "RemoveContainer" containerID="c5d320031c5314650cf936d492d9012062039ef119d0a1c04f3a620dc99aa3dd" Dec 03 13:32:39 crc kubenswrapper[4690]: I1203 13:32:39.060638 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" Dec 03 13:32:39 crc kubenswrapper[4690]: I1203 13:32:39.100357 4690 scope.go:117] "RemoveContainer" containerID="ff1781670859a363e94b9b97a46cffda2a5bc9a2910db781dda9118fe7dd5431" Dec 03 13:32:39 crc kubenswrapper[4690]: I1203 13:32:39.103047 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-gkh4f"] Dec 03 13:32:39 crc kubenswrapper[4690]: I1203 13:32:39.114944 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-gkh4f"] Dec 03 13:32:40 crc kubenswrapper[4690]: I1203 13:32:40.073226 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2rlzz" event={"ID":"88d2e02a-af82-49f0-858a-9e9fa3dc4985","Type":"ContainerStarted","Data":"c052f9f5ff262734453b7188da832580e8edd7fd7095639527e4e1ba04504bed"} Dec 03 13:32:40 crc kubenswrapper[4690]: I1203 13:32:40.107199 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-2rlzz" podStartSLOduration=3.7304855310000002 podStartE2EDuration="54.107170485s" podCreationTimestamp="2025-12-03 13:31:46 +0000 UTC" firstStartedPulling="2025-12-03 13:31:48.322964263 +0000 UTC m=+1354.303884696" lastFinishedPulling="2025-12-03 13:32:38.699649217 +0000 UTC m=+1404.680569650" observedRunningTime="2025-12-03 13:32:40.090418951 +0000 UTC m=+1406.071339404" watchObservedRunningTime="2025-12-03 13:32:40.107170485 +0000 UTC m=+1406.088090918" Dec 03 13:32:40 crc kubenswrapper[4690]: I1203 13:32:40.227773 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:40 crc kubenswrapper[4690]: I1203 13:32:40.227831 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:40 crc kubenswrapper[4690]: I1203 13:32:40.273505 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:40 crc kubenswrapper[4690]: I1203 13:32:40.282340 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:40 crc kubenswrapper[4690]: I1203 13:32:40.334553 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" path="/var/lib/kubelet/pods/cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b/volumes" Dec 03 13:32:41 crc kubenswrapper[4690]: I1203 13:32:41.093758 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:41 crc kubenswrapper[4690]: I1203 13:32:41.093808 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:41 crc kubenswrapper[4690]: I1203 13:32:41.225853 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 13:32:41 crc kubenswrapper[4690]: I1203 13:32:41.225920 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 13:32:41 crc kubenswrapper[4690]: I1203 13:32:41.268348 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 13:32:41 crc kubenswrapper[4690]: I1203 13:32:41.271915 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 13:32:42 crc kubenswrapper[4690]: I1203 13:32:42.103438 4690 generic.go:334] "Generic (PLEG): container finished" podID="29bf3e45-7939-4b48-923b-d0f28bf5c046" containerID="84d532dd714b5dbf5bc6ee00db6886ddf3ebe64dfec1465c5d61f5ac86193c30" exitCode=0 Dec 03 13:32:42 crc kubenswrapper[4690]: I1203 13:32:42.105229 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vcg6" event={"ID":"29bf3e45-7939-4b48-923b-d0f28bf5c046","Type":"ContainerDied","Data":"84d532dd714b5dbf5bc6ee00db6886ddf3ebe64dfec1465c5d61f5ac86193c30"} Dec 03 13:32:42 crc kubenswrapper[4690]: I1203 13:32:42.105265 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 13:32:42 crc kubenswrapper[4690]: I1203 13:32:42.106068 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 13:32:42 crc kubenswrapper[4690]: I1203 13:32:42.625110 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-gkh4f" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: i/o timeout" Dec 03 13:32:43 crc kubenswrapper[4690]: I1203 13:32:43.112910 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:32:43 crc kubenswrapper[4690]: I1203 13:32:43.113285 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:32:43 crc kubenswrapper[4690]: I1203 13:32:43.329722 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:43 crc kubenswrapper[4690]: I1203 13:32:43.481177 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 13:32:44 crc kubenswrapper[4690]: I1203 13:32:44.124368 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:32:44 crc kubenswrapper[4690]: I1203 13:32:44.124719 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:32:44 crc kubenswrapper[4690]: I1203 13:32:44.192667 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 13:32:44 crc kubenswrapper[4690]: I1203 13:32:44.195678 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.738807 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vcg6" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.855969 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bf3e45-7939-4b48-923b-d0f28bf5c046-logs\") pod \"29bf3e45-7939-4b48-923b-d0f28bf5c046\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.856069 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-combined-ca-bundle\") pod \"29bf3e45-7939-4b48-923b-d0f28bf5c046\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.856181 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-scripts\") pod \"29bf3e45-7939-4b48-923b-d0f28bf5c046\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.856219 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7rzg\" (UniqueName: \"kubernetes.io/projected/29bf3e45-7939-4b48-923b-d0f28bf5c046-kube-api-access-x7rzg\") pod \"29bf3e45-7939-4b48-923b-d0f28bf5c046\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.856315 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-config-data\") pod \"29bf3e45-7939-4b48-923b-d0f28bf5c046\" (UID: \"29bf3e45-7939-4b48-923b-d0f28bf5c046\") " Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.857782 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29bf3e45-7939-4b48-923b-d0f28bf5c046-logs" (OuterVolumeSpecName: "logs") pod "29bf3e45-7939-4b48-923b-d0f28bf5c046" (UID: "29bf3e45-7939-4b48-923b-d0f28bf5c046"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.862513 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29bf3e45-7939-4b48-923b-d0f28bf5c046-kube-api-access-x7rzg" (OuterVolumeSpecName: "kube-api-access-x7rzg") pod "29bf3e45-7939-4b48-923b-d0f28bf5c046" (UID: "29bf3e45-7939-4b48-923b-d0f28bf5c046"). InnerVolumeSpecName "kube-api-access-x7rzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.864161 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-scripts" (OuterVolumeSpecName: "scripts") pod "29bf3e45-7939-4b48-923b-d0f28bf5c046" (UID: "29bf3e45-7939-4b48-923b-d0f28bf5c046"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.893590 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29bf3e45-7939-4b48-923b-d0f28bf5c046" (UID: "29bf3e45-7939-4b48-923b-d0f28bf5c046"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.893687 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-config-data" (OuterVolumeSpecName: "config-data") pod "29bf3e45-7939-4b48-923b-d0f28bf5c046" (UID: "29bf3e45-7939-4b48-923b-d0f28bf5c046"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.958248 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bf3e45-7939-4b48-923b-d0f28bf5c046-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.958291 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.958304 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.958312 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7rzg\" (UniqueName: \"kubernetes.io/projected/29bf3e45-7939-4b48-923b-d0f28bf5c046-kube-api-access-x7rzg\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:45 crc kubenswrapper[4690]: I1203 13:32:45.958322 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bf3e45-7939-4b48-923b-d0f28bf5c046-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.145165 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vcg6" event={"ID":"29bf3e45-7939-4b48-923b-d0f28bf5c046","Type":"ContainerDied","Data":"3de4a911250866c5c93cc5ee2b044f75158f7189df676aa06fcabd9dbeb9626c"} Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.145550 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3de4a911250866c5c93cc5ee2b044f75158f7189df676aa06fcabd9dbeb9626c" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.146359 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vcg6" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.823641 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.823746 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.887605 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7f775c89b8-4tcrc"] Dec 03 13:32:46 crc kubenswrapper[4690]: E1203 13:32:46.888480 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29bf3e45-7939-4b48-923b-d0f28bf5c046" containerName="placement-db-sync" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.888651 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="29bf3e45-7939-4b48-923b-d0f28bf5c046" containerName="placement-db-sync" Dec 03 13:32:46 crc kubenswrapper[4690]: E1203 13:32:46.888762 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="init" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.888834 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="init" Dec 03 13:32:46 crc kubenswrapper[4690]: E1203 13:32:46.891113 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="dnsmasq-dns" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.891315 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="dnsmasq-dns" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.891856 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf4d9c9-d389-4aba-ad2a-f3e95f054f1b" containerName="dnsmasq-dns" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.891994 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="29bf3e45-7939-4b48-923b-d0f28bf5c046" containerName="placement-db-sync" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.893540 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.897853 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.897909 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.898101 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.898358 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.898405 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ckjs2" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.900561 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7f775c89b8-4tcrc"] Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.979944 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-public-tls-certs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.980047 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-combined-ca-bundle\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.980101 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-internal-tls-certs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.980137 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9a732dd-555c-4e51-a334-80dab3104192-logs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.980169 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-config-data\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.980206 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-scripts\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:46 crc kubenswrapper[4690]: I1203 13:32:46.980236 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdr2x\" (UniqueName: \"kubernetes.io/projected/f9a732dd-555c-4e51-a334-80dab3104192-kube-api-access-xdr2x\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.082158 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-config-data\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.082621 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-scripts\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.082754 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdr2x\" (UniqueName: \"kubernetes.io/projected/f9a732dd-555c-4e51-a334-80dab3104192-kube-api-access-xdr2x\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.082887 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-public-tls-certs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.083167 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-combined-ca-bundle\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.083346 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-internal-tls-certs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.083553 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9a732dd-555c-4e51-a334-80dab3104192-logs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.084206 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9a732dd-555c-4e51-a334-80dab3104192-logs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.092340 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-public-tls-certs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.092922 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-internal-tls-certs\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.102734 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-combined-ca-bundle\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.103216 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-scripts\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.103615 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9a732dd-555c-4e51-a334-80dab3104192-config-data\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.107280 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdr2x\" (UniqueName: \"kubernetes.io/projected/f9a732dd-555c-4e51-a334-80dab3104192-kube-api-access-xdr2x\") pod \"placement-7f775c89b8-4tcrc\" (UID: \"f9a732dd-555c-4e51-a334-80dab3104192\") " pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.219387 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:47 crc kubenswrapper[4690]: E1203 13:32:47.630441 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" Dec 03 13:32:47 crc kubenswrapper[4690]: I1203 13:32:47.724685 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7f775c89b8-4tcrc"] Dec 03 13:32:47 crc kubenswrapper[4690]: W1203 13:32:47.735427 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9a732dd_555c_4e51_a334_80dab3104192.slice/crio-ec18b8987cae164735dc3d274e17a85bfcf01bdc221659a6e9d7805c6b63feb5 WatchSource:0}: Error finding container ec18b8987cae164735dc3d274e17a85bfcf01bdc221659a6e9d7805c6b63feb5: Status 404 returned error can't find the container with id ec18b8987cae164735dc3d274e17a85bfcf01bdc221659a6e9d7805c6b63feb5 Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.179257 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"942f1c14-f12c-499f-9cd9-934bcdcfee99","Type":"ContainerStarted","Data":"afb76419d4add6b21d9a89f01f62975e7257689f5e8eff5df5a7ce85cc01ab48"} Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.179685 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.179517 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerName="proxy-httpd" containerID="cri-o://afb76419d4add6b21d9a89f01f62975e7257689f5e8eff5df5a7ce85cc01ab48" gracePeriod=30 Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.179396 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerName="ceilometer-notification-agent" containerID="cri-o://fe584c98cbe2ae53557fa5355768068e0600db371b4937efe754653e8ac46240" gracePeriod=30 Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.181835 4690 generic.go:334] "Generic (PLEG): container finished" podID="62230b01-b33e-4213-80c3-ecf6d5f995b1" containerID="be3917a1f502330f72af2766300a0420d1fbdd574129993d9051d6ac51c7b363" exitCode=0 Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.181943 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ndpp9" event={"ID":"62230b01-b33e-4213-80c3-ecf6d5f995b1","Type":"ContainerDied","Data":"be3917a1f502330f72af2766300a0420d1fbdd574129993d9051d6ac51c7b363"} Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.186656 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f775c89b8-4tcrc" event={"ID":"f9a732dd-555c-4e51-a334-80dab3104192","Type":"ContainerStarted","Data":"3ecedaf8af159d2b9516afb3bb762759fb415c4daf1624bdf8a565ce2d99bc86"} Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.186712 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f775c89b8-4tcrc" event={"ID":"f9a732dd-555c-4e51-a334-80dab3104192","Type":"ContainerStarted","Data":"b813c6e6b3de6bbc0de8500f3f96e94e175143bbaa1dc59eb6500aa24c881026"} Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.186725 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f775c89b8-4tcrc" event={"ID":"f9a732dd-555c-4e51-a334-80dab3104192","Type":"ContainerStarted","Data":"ec18b8987cae164735dc3d274e17a85bfcf01bdc221659a6e9d7805c6b63feb5"} Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.186813 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.186858 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:32:48 crc kubenswrapper[4690]: I1203 13:32:48.249210 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7f775c89b8-4tcrc" podStartSLOduration=2.249182702 podStartE2EDuration="2.249182702s" podCreationTimestamp="2025-12-03 13:32:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:48.237743706 +0000 UTC m=+1414.218664149" watchObservedRunningTime="2025-12-03 13:32:48.249182702 +0000 UTC m=+1414.230103135" Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.201152 4690 generic.go:334] "Generic (PLEG): container finished" podID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerID="afb76419d4add6b21d9a89f01f62975e7257689f5e8eff5df5a7ce85cc01ab48" exitCode=0 Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.201243 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"942f1c14-f12c-499f-9cd9-934bcdcfee99","Type":"ContainerDied","Data":"afb76419d4add6b21d9a89f01f62975e7257689f5e8eff5df5a7ce85cc01ab48"} Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.555199 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.643213 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-combined-ca-bundle\") pod \"62230b01-b33e-4213-80c3-ecf6d5f995b1\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.643291 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4smcb\" (UniqueName: \"kubernetes.io/projected/62230b01-b33e-4213-80c3-ecf6d5f995b1-kube-api-access-4smcb\") pod \"62230b01-b33e-4213-80c3-ecf6d5f995b1\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.643334 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-db-sync-config-data\") pod \"62230b01-b33e-4213-80c3-ecf6d5f995b1\" (UID: \"62230b01-b33e-4213-80c3-ecf6d5f995b1\") " Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.650047 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62230b01-b33e-4213-80c3-ecf6d5f995b1-kube-api-access-4smcb" (OuterVolumeSpecName: "kube-api-access-4smcb") pod "62230b01-b33e-4213-80c3-ecf6d5f995b1" (UID: "62230b01-b33e-4213-80c3-ecf6d5f995b1"). InnerVolumeSpecName "kube-api-access-4smcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.653408 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "62230b01-b33e-4213-80c3-ecf6d5f995b1" (UID: "62230b01-b33e-4213-80c3-ecf6d5f995b1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.688060 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62230b01-b33e-4213-80c3-ecf6d5f995b1" (UID: "62230b01-b33e-4213-80c3-ecf6d5f995b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.745677 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.745718 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4smcb\" (UniqueName: \"kubernetes.io/projected/62230b01-b33e-4213-80c3-ecf6d5f995b1-kube-api-access-4smcb\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:49 crc kubenswrapper[4690]: I1203 13:32:49.745734 4690 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62230b01-b33e-4213-80c3-ecf6d5f995b1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.212557 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ndpp9" event={"ID":"62230b01-b33e-4213-80c3-ecf6d5f995b1","Type":"ContainerDied","Data":"9d23ee71098c333ecf04b12d677c1953379bd96e598dc886319183111d92c0ae"} Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.212611 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d23ee71098c333ecf04b12d677c1953379bd96e598dc886319183111d92c0ae" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.212638 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ndpp9" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.479046 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7c7c4b9c77-f54ln"] Dec 03 13:32:50 crc kubenswrapper[4690]: E1203 13:32:50.479550 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62230b01-b33e-4213-80c3-ecf6d5f995b1" containerName="barbican-db-sync" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.479568 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="62230b01-b33e-4213-80c3-ecf6d5f995b1" containerName="barbican-db-sync" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.479743 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="62230b01-b33e-4213-80c3-ecf6d5f995b1" containerName="barbican-db-sync" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.481501 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.484976 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-22rqt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.493036 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.495172 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.502654 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7c7c4b9c77-f54ln"] Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.512030 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-98c9dfb78-qjztd"] Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.513730 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.518454 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.534905 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-98c9dfb78-qjztd"] Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.567462 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqncf\" (UniqueName: \"kubernetes.io/projected/f7a19182-b617-47e3-bc7f-db372257b984-kube-api-access-jqncf\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.567559 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-combined-ca-bundle\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.567598 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-config-data-custom\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.567652 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a19182-b617-47e3-bc7f-db372257b984-logs\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.567685 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-config-data\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.621655 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-l66qt"] Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.623912 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.631267 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-l66qt"] Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674216 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghkdm\" (UniqueName: \"kubernetes.io/projected/790eb625-6e7d-4467-afef-4073973889b2-kube-api-access-ghkdm\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674278 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/790eb625-6e7d-4467-afef-4073973889b2-logs\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674313 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqncf\" (UniqueName: \"kubernetes.io/projected/f7a19182-b617-47e3-bc7f-db372257b984-kube-api-access-jqncf\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674392 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-combined-ca-bundle\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674431 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-config-data-custom\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674461 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-config-data\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674502 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-config-data-custom\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674542 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a19182-b617-47e3-bc7f-db372257b984-logs\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674573 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-config-data\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.674629 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-combined-ca-bundle\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.675454 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a19182-b617-47e3-bc7f-db372257b984-logs\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.688023 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-config-data\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.695551 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-combined-ca-bundle\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.700324 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqncf\" (UniqueName: \"kubernetes.io/projected/f7a19182-b617-47e3-bc7f-db372257b984-kube-api-access-jqncf\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.706484 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7a19182-b617-47e3-bc7f-db372257b984-config-data-custom\") pod \"barbican-worker-7c7c4b9c77-f54ln\" (UID: \"f7a19182-b617-47e3-bc7f-db372257b984\") " pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.733492 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8bf94c858-npzt6"] Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.737079 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.743060 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.778839 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.778928 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghkdm\" (UniqueName: \"kubernetes.io/projected/790eb625-6e7d-4467-afef-4073973889b2-kube-api-access-ghkdm\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.778958 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnh6j\" (UniqueName: \"kubernetes.io/projected/14453e43-831f-4ea4-9a7f-57621e3999a4-kube-api-access-rnh6j\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.778986 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/790eb625-6e7d-4467-afef-4073973889b2-logs\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.779048 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.779097 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-config-data\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.779131 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-config-data-custom\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.779189 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-combined-ca-bundle\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.779223 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-config\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.779260 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.779304 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.780218 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/790eb625-6e7d-4467-afef-4073973889b2-logs\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.780473 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8bf94c858-npzt6"] Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.808036 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-combined-ca-bundle\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.808383 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-config-data-custom\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.808725 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c7c4b9c77-f54ln" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.809096 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/790eb625-6e7d-4467-afef-4073973889b2-config-data\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.813491 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghkdm\" (UniqueName: \"kubernetes.io/projected/790eb625-6e7d-4467-afef-4073973889b2-kube-api-access-ghkdm\") pod \"barbican-keystone-listener-98c9dfb78-qjztd\" (UID: \"790eb625-6e7d-4467-afef-4073973889b2\") " pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.854979 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.880791 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-config\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.880849 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.880909 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9m6k\" (UniqueName: \"kubernetes.io/projected/a683ac29-783f-4210-97dc-42b0deff462d-kube-api-access-c9m6k\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.880935 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.880960 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data-custom\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.880977 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a683ac29-783f-4210-97dc-42b0deff462d-logs\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.881004 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.883816 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnh6j\" (UniqueName: \"kubernetes.io/projected/14453e43-831f-4ea4-9a7f-57621e3999a4-kube-api-access-rnh6j\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.883896 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.884037 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.884124 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-combined-ca-bundle\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.885949 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-config\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.886353 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.886854 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.890388 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.891208 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.913750 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnh6j\" (UniqueName: \"kubernetes.io/projected/14453e43-831f-4ea4-9a7f-57621e3999a4-kube-api-access-rnh6j\") pod \"dnsmasq-dns-586bdc5f9-l66qt\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.970325 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.989784 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9m6k\" (UniqueName: \"kubernetes.io/projected/a683ac29-783f-4210-97dc-42b0deff462d-kube-api-access-c9m6k\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.990362 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data-custom\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.990388 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a683ac29-783f-4210-97dc-42b0deff462d-logs\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.990443 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.990521 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-combined-ca-bundle\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:50 crc kubenswrapper[4690]: I1203 13:32:50.991146 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a683ac29-783f-4210-97dc-42b0deff462d-logs\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:50.995523 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data-custom\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:50.998833 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-combined-ca-bundle\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:50.999768 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:51.010422 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9m6k\" (UniqueName: \"kubernetes.io/projected/a683ac29-783f-4210-97dc-42b0deff462d-kube-api-access-c9m6k\") pod \"barbican-api-8bf94c858-npzt6\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:51.243846 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:51.428181 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7c7c4b9c77-f54ln"] Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:51.442501 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-98c9dfb78-qjztd"] Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:51.703984 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-l66qt"] Dec 03 13:32:51 crc kubenswrapper[4690]: W1203 13:32:51.845932 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda683ac29_783f_4210_97dc_42b0deff462d.slice/crio-48ac755eb751d3ffd95c6ab82d52100f2002246baa335c37305d8a98023b8608 WatchSource:0}: Error finding container 48ac755eb751d3ffd95c6ab82d52100f2002246baa335c37305d8a98023b8608: Status 404 returned error can't find the container with id 48ac755eb751d3ffd95c6ab82d52100f2002246baa335c37305d8a98023b8608 Dec 03 13:32:51 crc kubenswrapper[4690]: I1203 13:32:51.849613 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8bf94c858-npzt6"] Dec 03 13:32:52 crc kubenswrapper[4690]: I1203 13:32:52.239085 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" event={"ID":"14453e43-831f-4ea4-9a7f-57621e3999a4","Type":"ContainerStarted","Data":"dbd56c23b40d664452b75f43ff085eb3ab84597759cf3f2d088d91805c35ef7d"} Dec 03 13:32:52 crc kubenswrapper[4690]: I1203 13:32:52.239153 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" event={"ID":"14453e43-831f-4ea4-9a7f-57621e3999a4","Type":"ContainerStarted","Data":"915f78d71448b06513375144c0f2219d6cf16634b09897024f6c79ad11c7c4df"} Dec 03 13:32:52 crc kubenswrapper[4690]: I1203 13:32:52.241074 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8bf94c858-npzt6" event={"ID":"a683ac29-783f-4210-97dc-42b0deff462d","Type":"ContainerStarted","Data":"48ac755eb751d3ffd95c6ab82d52100f2002246baa335c37305d8a98023b8608"} Dec 03 13:32:52 crc kubenswrapper[4690]: I1203 13:32:52.242634 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c7c4b9c77-f54ln" event={"ID":"f7a19182-b617-47e3-bc7f-db372257b984","Type":"ContainerStarted","Data":"fe9b3c5d467c6c55a65f0df91442277eee6d23088490ae6fb44031c7d1c1cb03"} Dec 03 13:32:52 crc kubenswrapper[4690]: I1203 13:32:52.243965 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" event={"ID":"790eb625-6e7d-4467-afef-4073973889b2","Type":"ContainerStarted","Data":"4bca8ce0e8bad1302df047fc558da48e90a792ef095ab606db1beddd97aae45f"} Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.258211 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8bf94c858-npzt6" event={"ID":"a683ac29-783f-4210-97dc-42b0deff462d","Type":"ContainerStarted","Data":"1ca6d398d142cc13c7aa5bb45a4f7f19ed37784935f0174787815fd5f45b0ebe"} Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.264302 4690 generic.go:334] "Generic (PLEG): container finished" podID="14453e43-831f-4ea4-9a7f-57621e3999a4" containerID="dbd56c23b40d664452b75f43ff085eb3ab84597759cf3f2d088d91805c35ef7d" exitCode=0 Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.264351 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" event={"ID":"14453e43-831f-4ea4-9a7f-57621e3999a4","Type":"ContainerDied","Data":"dbd56c23b40d664452b75f43ff085eb3ab84597759cf3f2d088d91805c35ef7d"} Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.594055 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-d8d5ff786-rbh82"] Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.599499 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.604385 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.604652 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.615098 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d8d5ff786-rbh82"] Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.759731 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-logs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.760125 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-config-data-custom\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.760157 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-combined-ca-bundle\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.760238 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-config-data\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.760296 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-public-tls-certs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.760320 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb7rj\" (UniqueName: \"kubernetes.io/projected/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-kube-api-access-zb7rj\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.760381 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-internal-tls-certs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.862165 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-config-data\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.862280 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-public-tls-certs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.862308 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb7rj\" (UniqueName: \"kubernetes.io/projected/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-kube-api-access-zb7rj\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.862352 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-internal-tls-certs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.862476 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-logs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.862508 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-config-data-custom\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.862533 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-combined-ca-bundle\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.863528 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-logs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.868548 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-config-data-custom\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.872369 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-internal-tls-certs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.876486 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-public-tls-certs\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.882770 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-combined-ca-bundle\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.887079 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-config-data\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.892462 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb7rj\" (UniqueName: \"kubernetes.io/projected/1a74f0f7-bcc8-483f-ad5a-2df09e7f462d-kube-api-access-zb7rj\") pod \"barbican-api-d8d5ff786-rbh82\" (UID: \"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d\") " pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:53 crc kubenswrapper[4690]: I1203 13:32:53.971261 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.191572 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-788df6b698-7gkrp" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.289898 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8bf94c858-npzt6" event={"ID":"a683ac29-783f-4210-97dc-42b0deff462d","Type":"ContainerStarted","Data":"f164137970f4783e844b6f624ab3958fdf03cf267f74f31ce3f5db4928eb0a9c"} Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.290299 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.290363 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.293805 4690 generic.go:334] "Generic (PLEG): container finished" podID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerID="fe584c98cbe2ae53557fa5355768068e0600db371b4937efe754653e8ac46240" exitCode=0 Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.293911 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"942f1c14-f12c-499f-9cd9-934bcdcfee99","Type":"ContainerDied","Data":"fe584c98cbe2ae53557fa5355768068e0600db371b4937efe754653e8ac46240"} Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.303672 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" event={"ID":"14453e43-831f-4ea4-9a7f-57621e3999a4","Type":"ContainerStarted","Data":"96163939beeee95229d9a3343b458be1a7d7e08afe9b3b9a31a7899ea09a9d1e"} Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.313733 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8bf94c858-npzt6" podStartSLOduration=4.313716055 podStartE2EDuration="4.313716055s" podCreationTimestamp="2025-12-03 13:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:54.307186336 +0000 UTC m=+1420.288106789" watchObservedRunningTime="2025-12-03 13:32:54.313716055 +0000 UTC m=+1420.294636488" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.509474 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d8d5ff786-rbh82"] Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.787143 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.905371 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-sg-core-conf-yaml\") pod \"942f1c14-f12c-499f-9cd9-934bcdcfee99\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.905490 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-log-httpd\") pod \"942f1c14-f12c-499f-9cd9-934bcdcfee99\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.905530 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-run-httpd\") pod \"942f1c14-f12c-499f-9cd9-934bcdcfee99\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.905569 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-config-data\") pod \"942f1c14-f12c-499f-9cd9-934bcdcfee99\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.905733 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-combined-ca-bundle\") pod \"942f1c14-f12c-499f-9cd9-934bcdcfee99\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.905848 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-scripts\") pod \"942f1c14-f12c-499f-9cd9-934bcdcfee99\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.905900 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mp8r\" (UniqueName: \"kubernetes.io/projected/942f1c14-f12c-499f-9cd9-934bcdcfee99-kube-api-access-7mp8r\") pod \"942f1c14-f12c-499f-9cd9-934bcdcfee99\" (UID: \"942f1c14-f12c-499f-9cd9-934bcdcfee99\") " Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.906035 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "942f1c14-f12c-499f-9cd9-934bcdcfee99" (UID: "942f1c14-f12c-499f-9cd9-934bcdcfee99"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.906375 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "942f1c14-f12c-499f-9cd9-934bcdcfee99" (UID: "942f1c14-f12c-499f-9cd9-934bcdcfee99"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.906455 4690 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.906478 4690 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/942f1c14-f12c-499f-9cd9-934bcdcfee99-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.912808 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/942f1c14-f12c-499f-9cd9-934bcdcfee99-kube-api-access-7mp8r" (OuterVolumeSpecName: "kube-api-access-7mp8r") pod "942f1c14-f12c-499f-9cd9-934bcdcfee99" (UID: "942f1c14-f12c-499f-9cd9-934bcdcfee99"). InnerVolumeSpecName "kube-api-access-7mp8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.921596 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-scripts" (OuterVolumeSpecName: "scripts") pod "942f1c14-f12c-499f-9cd9-934bcdcfee99" (UID: "942f1c14-f12c-499f-9cd9-934bcdcfee99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.921768 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "942f1c14-f12c-499f-9cd9-934bcdcfee99" (UID: "942f1c14-f12c-499f-9cd9-934bcdcfee99"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.960680 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "942f1c14-f12c-499f-9cd9-934bcdcfee99" (UID: "942f1c14-f12c-499f-9cd9-934bcdcfee99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:54 crc kubenswrapper[4690]: I1203 13:32:54.994689 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-config-data" (OuterVolumeSpecName: "config-data") pod "942f1c14-f12c-499f-9cd9-934bcdcfee99" (UID: "942f1c14-f12c-499f-9cd9-934bcdcfee99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.008618 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.008832 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.008850 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.008861 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mp8r\" (UniqueName: \"kubernetes.io/projected/942f1c14-f12c-499f-9cd9-934bcdcfee99-kube-api-access-7mp8r\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.008907 4690 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/942f1c14-f12c-499f-9cd9-934bcdcfee99-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.314771 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d8d5ff786-rbh82" event={"ID":"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d","Type":"ContainerStarted","Data":"eed56482cb00c02cf55bb5aae4f6a0c1de67a116bccb6983d34b25d312e46a2f"} Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.317657 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"942f1c14-f12c-499f-9cd9-934bcdcfee99","Type":"ContainerDied","Data":"77a82477e445c6a37bb46a1f9d80462479f3139cb4d8fa063fd3adc0cbe79976"} Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.317697 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.317719 4690 scope.go:117] "RemoveContainer" containerID="afb76419d4add6b21d9a89f01f62975e7257689f5e8eff5df5a7ce85cc01ab48" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.351752 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" podStartSLOduration=5.351734659 podStartE2EDuration="5.351734659s" podCreationTimestamp="2025-12-03 13:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:32:55.348486445 +0000 UTC m=+1421.329406898" watchObservedRunningTime="2025-12-03 13:32:55.351734659 +0000 UTC m=+1421.332655092" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.425439 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.437822 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.450523 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:32:55 crc kubenswrapper[4690]: E1203 13:32:55.451216 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerName="ceilometer-notification-agent" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.451280 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerName="ceilometer-notification-agent" Dec 03 13:32:55 crc kubenswrapper[4690]: E1203 13:32:55.451301 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerName="proxy-httpd" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.451307 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerName="proxy-httpd" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.451669 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerName="proxy-httpd" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.451689 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" containerName="ceilometer-notification-agent" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.453550 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.456668 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.456931 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.474300 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.623182 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-scripts\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.623608 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.623935 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdzj7\" (UniqueName: \"kubernetes.io/projected/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-kube-api-access-wdzj7\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.624090 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-config-data\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.624319 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-log-httpd\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.624497 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-run-httpd\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.624724 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.727261 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.727333 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-scripts\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.727506 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.727626 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdzj7\" (UniqueName: \"kubernetes.io/projected/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-kube-api-access-wdzj7\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.727652 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-config-data\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.727711 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-log-httpd\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.727746 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-run-httpd\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.729001 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-log-httpd\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.729049 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-run-httpd\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.734536 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.734661 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.737659 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-scripts\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.747297 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-config-data\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.759350 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdzj7\" (UniqueName: \"kubernetes.io/projected/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-kube-api-access-wdzj7\") pod \"ceilometer-0\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.775779 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:32:55 crc kubenswrapper[4690]: I1203 13:32:55.970946 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:32:56 crc kubenswrapper[4690]: I1203 13:32:56.327574 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="942f1c14-f12c-499f-9cd9-934bcdcfee99" path="/var/lib/kubelet/pods/942f1c14-f12c-499f-9cd9-934bcdcfee99/volumes" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.244047 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.247244 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.251361 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.251361 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-wqxjd" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.251426 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.260358 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.263246 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.263299 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.263542 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdv6p\" (UniqueName: \"kubernetes.io/projected/c169e895-945b-4cac-baee-75d5e38d6ba7-kube-api-access-rdv6p\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.263723 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config-secret\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.365572 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config-secret\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.365791 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.365820 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.365896 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdv6p\" (UniqueName: \"kubernetes.io/projected/c169e895-945b-4cac-baee-75d5e38d6ba7-kube-api-access-rdv6p\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.367709 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.374828 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config-secret\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.376835 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.388191 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdv6p\" (UniqueName: \"kubernetes.io/projected/c169e895-945b-4cac-baee-75d5e38d6ba7-kube-api-access-rdv6p\") pod \"openstackclient\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.586112 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.784598 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.793263 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.824000 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.825645 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.837659 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.875136 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f47b845f-82f3-44b9-966d-eb3076f05362-openstack-config-secret\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.875603 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f47b845f-82f3-44b9-966d-eb3076f05362-openstack-config\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.875666 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f47b845f-82f3-44b9-966d-eb3076f05362-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.875739 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7hcq\" (UniqueName: \"kubernetes.io/projected/f47b845f-82f3-44b9-966d-eb3076f05362-kube-api-access-p7hcq\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.977667 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f47b845f-82f3-44b9-966d-eb3076f05362-openstack-config-secret\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.977770 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f47b845f-82f3-44b9-966d-eb3076f05362-openstack-config\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.977923 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f47b845f-82f3-44b9-966d-eb3076f05362-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.978042 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7hcq\" (UniqueName: \"kubernetes.io/projected/f47b845f-82f3-44b9-966d-eb3076f05362-kube-api-access-p7hcq\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.979403 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f47b845f-82f3-44b9-966d-eb3076f05362-openstack-config\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.986370 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f47b845f-82f3-44b9-966d-eb3076f05362-openstack-config-secret\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:57 crc kubenswrapper[4690]: I1203 13:32:57.987183 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f47b845f-82f3-44b9-966d-eb3076f05362-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:58 crc kubenswrapper[4690]: I1203 13:32:58.003491 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7hcq\" (UniqueName: \"kubernetes.io/projected/f47b845f-82f3-44b9-966d-eb3076f05362-kube-api-access-p7hcq\") pod \"openstackclient\" (UID: \"f47b845f-82f3-44b9-966d-eb3076f05362\") " pod="openstack/openstackclient" Dec 03 13:32:58 crc kubenswrapper[4690]: I1203 13:32:58.151170 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:32:58 crc kubenswrapper[4690]: I1203 13:32:58.934080 4690 scope.go:117] "RemoveContainer" containerID="fe584c98cbe2ae53557fa5355768068e0600db371b4937efe754653e8ac46240" Dec 03 13:32:59 crc kubenswrapper[4690]: I1203 13:32:59.482064 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 13:32:59 crc kubenswrapper[4690]: W1203 13:32:59.486094 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf47b845f_82f3_44b9_966d_eb3076f05362.slice/crio-49f41ae523b407a6dfa396071355d0a1a7963fc2a606e81ff8d5e8de1de79c23 WatchSource:0}: Error finding container 49f41ae523b407a6dfa396071355d0a1a7963fc2a606e81ff8d5e8de1de79c23: Status 404 returned error can't find the container with id 49f41ae523b407a6dfa396071355d0a1a7963fc2a606e81ff8d5e8de1de79c23 Dec 03 13:32:59 crc kubenswrapper[4690]: E1203 13:32:59.543818 4690 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 13:32:59 crc kubenswrapper[4690]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_c169e895-945b-4cac-baee-75d5e38d6ba7_0(bf9b355df1d1f97d9c6209cb2e6f5152707302f9a5dd1e98b0e5a41b051c408a): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bf9b355df1d1f97d9c6209cb2e6f5152707302f9a5dd1e98b0e5a41b051c408a" Netns:"/var/run/netns/d3becb59-a9c5-4195-bbc8-ed29259cad5f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=bf9b355df1d1f97d9c6209cb2e6f5152707302f9a5dd1e98b0e5a41b051c408a;K8S_POD_UID=c169e895-945b-4cac-baee-75d5e38d6ba7" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/c169e895-945b-4cac-baee-75d5e38d6ba7]: expected pod UID "c169e895-945b-4cac-baee-75d5e38d6ba7" but got "f47b845f-82f3-44b9-966d-eb3076f05362" from Kube API Dec 03 13:32:59 crc kubenswrapper[4690]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 13:32:59 crc kubenswrapper[4690]: > Dec 03 13:32:59 crc kubenswrapper[4690]: E1203 13:32:59.544198 4690 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 13:32:59 crc kubenswrapper[4690]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_c169e895-945b-4cac-baee-75d5e38d6ba7_0(bf9b355df1d1f97d9c6209cb2e6f5152707302f9a5dd1e98b0e5a41b051c408a): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bf9b355df1d1f97d9c6209cb2e6f5152707302f9a5dd1e98b0e5a41b051c408a" Netns:"/var/run/netns/d3becb59-a9c5-4195-bbc8-ed29259cad5f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=bf9b355df1d1f97d9c6209cb2e6f5152707302f9a5dd1e98b0e5a41b051c408a;K8S_POD_UID=c169e895-945b-4cac-baee-75d5e38d6ba7" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/c169e895-945b-4cac-baee-75d5e38d6ba7]: expected pod UID "c169e895-945b-4cac-baee-75d5e38d6ba7" but got "f47b845f-82f3-44b9-966d-eb3076f05362" from Kube API Dec 03 13:32:59 crc kubenswrapper[4690]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 13:32:59 crc kubenswrapper[4690]: > pod="openstack/openstackclient" Dec 03 13:32:59 crc kubenswrapper[4690]: I1203 13:32:59.570340 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:32:59 crc kubenswrapper[4690]: W1203 13:32:59.573609 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode23c2dcc_ec98_49a4_92e9_db075d8f6f42.slice/crio-8c342cd85095f4d239f03185c9c6245d140cb5aa22c6d1050fcdfd9748ae4fb5 WatchSource:0}: Error finding container 8c342cd85095f4d239f03185c9c6245d140cb5aa22c6d1050fcdfd9748ae4fb5: Status 404 returned error can't find the container with id 8c342cd85095f4d239f03185c9c6245d140cb5aa22c6d1050fcdfd9748ae4fb5 Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.377752 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f47b845f-82f3-44b9-966d-eb3076f05362","Type":"ContainerStarted","Data":"49f41ae523b407a6dfa396071355d0a1a7963fc2a606e81ff8d5e8de1de79c23"} Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.379183 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerStarted","Data":"8c342cd85095f4d239f03185c9c6245d140cb5aa22c6d1050fcdfd9748ae4fb5"} Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.381817 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d8d5ff786-rbh82" event={"ID":"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d","Type":"ContainerStarted","Data":"170c414576a6190bad3cf4efca0713000fdbc8e5857ea16f5b6bdaed54bc0657"} Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.381895 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.402664 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.408516 4690 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c169e895-945b-4cac-baee-75d5e38d6ba7" podUID="f47b845f-82f3-44b9-966d-eb3076f05362" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.437532 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config\") pod \"c169e895-945b-4cac-baee-75d5e38d6ba7\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.437689 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdv6p\" (UniqueName: \"kubernetes.io/projected/c169e895-945b-4cac-baee-75d5e38d6ba7-kube-api-access-rdv6p\") pod \"c169e895-945b-4cac-baee-75d5e38d6ba7\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.438612 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-combined-ca-bundle\") pod \"c169e895-945b-4cac-baee-75d5e38d6ba7\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.438751 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config-secret\") pod \"c169e895-945b-4cac-baee-75d5e38d6ba7\" (UID: \"c169e895-945b-4cac-baee-75d5e38d6ba7\") " Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.438798 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c169e895-945b-4cac-baee-75d5e38d6ba7" (UID: "c169e895-945b-4cac-baee-75d5e38d6ba7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.440261 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.446935 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c169e895-945b-4cac-baee-75d5e38d6ba7-kube-api-access-rdv6p" (OuterVolumeSpecName: "kube-api-access-rdv6p") pod "c169e895-945b-4cac-baee-75d5e38d6ba7" (UID: "c169e895-945b-4cac-baee-75d5e38d6ba7"). InnerVolumeSpecName "kube-api-access-rdv6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.446980 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c169e895-945b-4cac-baee-75d5e38d6ba7" (UID: "c169e895-945b-4cac-baee-75d5e38d6ba7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.449451 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c169e895-945b-4cac-baee-75d5e38d6ba7" (UID: "c169e895-945b-4cac-baee-75d5e38d6ba7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.542604 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.542659 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c169e895-945b-4cac-baee-75d5e38d6ba7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.542677 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdv6p\" (UniqueName: \"kubernetes.io/projected/c169e895-945b-4cac-baee-75d5e38d6ba7-kube-api-access-rdv6p\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:00 crc kubenswrapper[4690]: I1203 13:33:00.972096 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:33:01 crc kubenswrapper[4690]: I1203 13:33:01.059515 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-j89sv"] Dec 03 13:33:01 crc kubenswrapper[4690]: I1203 13:33:01.059838 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" podUID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerName="dnsmasq-dns" containerID="cri-o://c0f7708e37c77380e77a4e58092d11da4ff74a489c99344b91225a9976f62aa1" gracePeriod=10 Dec 03 13:33:01 crc kubenswrapper[4690]: I1203 13:33:01.391338 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:33:01 crc kubenswrapper[4690]: I1203 13:33:01.406961 4690 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c169e895-945b-4cac-baee-75d5e38d6ba7" podUID="f47b845f-82f3-44b9-966d-eb3076f05362" Dec 03 13:33:02 crc kubenswrapper[4690]: I1203 13:33:02.287196 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-8bf94c858-npzt6" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:33:02 crc kubenswrapper[4690]: I1203 13:33:02.340590 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c169e895-945b-4cac-baee-75d5e38d6ba7" path="/var/lib/kubelet/pods/c169e895-945b-4cac-baee-75d5e38d6ba7/volumes" Dec 03 13:33:02 crc kubenswrapper[4690]: I1203 13:33:02.415819 4690 generic.go:334] "Generic (PLEG): container finished" podID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerID="c0f7708e37c77380e77a4e58092d11da4ff74a489c99344b91225a9976f62aa1" exitCode=0 Dec 03 13:33:02 crc kubenswrapper[4690]: I1203 13:33:02.415906 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" event={"ID":"c8fd9d7d-4064-48c5-9a0a-cabf1de25647","Type":"ContainerDied","Data":"c0f7708e37c77380e77a4e58092d11da4ff74a489c99344b91225a9976f62aa1"} Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.155170 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.167126 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.289660 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" podUID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.156:5353: connect: connection refused" Dec 03 13:33:03 crc kubenswrapper[4690]: E1203 13:33:03.511351 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified" Dec 03 13:33:03 crc kubenswrapper[4690]: E1203 13:33:03.511568 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-worker-log,Image:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,Command:[/usr/bin/dumb-init],Args:[--single-child -- /usr/bin/tail -n+1 -F /var/log/barbican/barbican-worker.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5fdhf7h5b8h7bh94hd4h579h56ch57dh684h5cdhd6h7dh5bdh544h67bh677h65ch56h95h66bh75h5cbh7hf8h58h59bh569h54dh4h5d8h9q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/barbican,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jqncf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-worker-7c7c4b9c77-f54ln_openstack(f7a19182-b617-47e3-bc7f-db372257b984): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:33:03 crc kubenswrapper[4690]: E1203 13:33:03.678184 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"barbican-worker-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"barbican-worker\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified\\\"\"]" pod="openstack/barbican-worker-7c7c4b9c77-f54ln" podUID="f7a19182-b617-47e3-bc7f-db372257b984" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.805356 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.825111 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-svc\") pod \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.825249 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slnfw\" (UniqueName: \"kubernetes.io/projected/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-kube-api-access-slnfw\") pod \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.825327 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-config\") pod \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.825401 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-nb\") pod \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.825448 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-sb\") pod \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.825506 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-swift-storage-0\") pod \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\" (UID: \"c8fd9d7d-4064-48c5-9a0a-cabf1de25647\") " Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.847118 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-kube-api-access-slnfw" (OuterVolumeSpecName: "kube-api-access-slnfw") pod "c8fd9d7d-4064-48c5-9a0a-cabf1de25647" (UID: "c8fd9d7d-4064-48c5-9a0a-cabf1de25647"). InnerVolumeSpecName "kube-api-access-slnfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.929615 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slnfw\" (UniqueName: \"kubernetes.io/projected/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-kube-api-access-slnfw\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.959321 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c8fd9d7d-4064-48c5-9a0a-cabf1de25647" (UID: "c8fd9d7d-4064-48c5-9a0a-cabf1de25647"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.962147 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c8fd9d7d-4064-48c5-9a0a-cabf1de25647" (UID: "c8fd9d7d-4064-48c5-9a0a-cabf1de25647"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.976167 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c8fd9d7d-4064-48c5-9a0a-cabf1de25647" (UID: "c8fd9d7d-4064-48c5-9a0a-cabf1de25647"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:03 crc kubenswrapper[4690]: I1203 13:33:03.989514 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-config" (OuterVolumeSpecName: "config") pod "c8fd9d7d-4064-48c5-9a0a-cabf1de25647" (UID: "c8fd9d7d-4064-48c5-9a0a-cabf1de25647"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.003135 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c8fd9d7d-4064-48c5-9a0a-cabf1de25647" (UID: "c8fd9d7d-4064-48c5-9a0a-cabf1de25647"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.032805 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.032843 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.032853 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.032876 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.032885 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c8fd9d7d-4064-48c5-9a0a-cabf1de25647-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.438713 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.438994 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-j89sv" event={"ID":"c8fd9d7d-4064-48c5-9a0a-cabf1de25647","Type":"ContainerDied","Data":"5e945d883d0dd47bc0c38178bb170b27117a0bac07d6efe0aeca799cc21d4884"} Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.439063 4690 scope.go:117] "RemoveContainer" containerID="c0f7708e37c77380e77a4e58092d11da4ff74a489c99344b91225a9976f62aa1" Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.485994 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-j89sv"] Dec 03 13:33:04 crc kubenswrapper[4690]: I1203 13:33:04.494323 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-j89sv"] Dec 03 13:33:05 crc kubenswrapper[4690]: I1203 13:33:05.451125 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d8d5ff786-rbh82" event={"ID":"1a74f0f7-bcc8-483f-ad5a-2df09e7f462d","Type":"ContainerStarted","Data":"c855459e13e454655fa64deb2387a00bb3fed3221960e4941130d7d33f02ca3c"} Dec 03 13:33:05 crc kubenswrapper[4690]: I1203 13:33:05.454995 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:33:05 crc kubenswrapper[4690]: I1203 13:33:05.455023 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:33:05 crc kubenswrapper[4690]: I1203 13:33:05.484327 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-d8d5ff786-rbh82" podStartSLOduration=12.484301466 podStartE2EDuration="12.484301466s" podCreationTimestamp="2025-12-03 13:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:05.476105314 +0000 UTC m=+1431.457025747" watchObservedRunningTime="2025-12-03 13:33:05.484301466 +0000 UTC m=+1431.465221899" Dec 03 13:33:05 crc kubenswrapper[4690]: E1203 13:33:05.616733 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"barbican-worker-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified\\\"\", failed to \"StartContainer\" for \"barbican-worker\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified\\\"\"]" pod="openstack/barbican-worker-7c7c4b9c77-f54ln" podUID="f7a19182-b617-47e3-bc7f-db372257b984" Dec 03 13:33:05 crc kubenswrapper[4690]: I1203 13:33:05.616724 4690 scope.go:117] "RemoveContainer" containerID="1786ab286f9279ade15677762d98310b86563a2576124c5135af41965bd238b7" Dec 03 13:33:06 crc kubenswrapper[4690]: I1203 13:33:06.327784 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" path="/var/lib/kubelet/pods/c8fd9d7d-4064-48c5-9a0a-cabf1de25647/volumes" Dec 03 13:33:06 crc kubenswrapper[4690]: I1203 13:33:06.467569 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerStarted","Data":"294b0b57fa1be9e729837ce3dda9c0dff30aab373239cb0aacc2f8d55c41747c"} Dec 03 13:33:06 crc kubenswrapper[4690]: I1203 13:33:06.472064 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" event={"ID":"790eb625-6e7d-4467-afef-4073973889b2","Type":"ContainerStarted","Data":"81e8c879267c148f03d559eb1c5c179cf80875c85345e5aeff14016d1f712a61"} Dec 03 13:33:06 crc kubenswrapper[4690]: I1203 13:33:06.472109 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" event={"ID":"790eb625-6e7d-4467-afef-4073973889b2","Type":"ContainerStarted","Data":"d780820ef5fde11d262b7bae4e5ff2d5d7d9ec937fc000e955560c696adf13ab"} Dec 03 13:33:06 crc kubenswrapper[4690]: I1203 13:33:06.511716 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-98c9dfb78-qjztd" podStartSLOduration=2.3562924450000002 podStartE2EDuration="16.511567011s" podCreationTimestamp="2025-12-03 13:32:50 +0000 UTC" firstStartedPulling="2025-12-03 13:32:51.463135161 +0000 UTC m=+1417.444055604" lastFinishedPulling="2025-12-03 13:33:05.618409737 +0000 UTC m=+1431.599330170" observedRunningTime="2025-12-03 13:33:06.492064216 +0000 UTC m=+1432.472984669" watchObservedRunningTime="2025-12-03 13:33:06.511567011 +0000 UTC m=+1432.492487454" Dec 03 13:33:07 crc kubenswrapper[4690]: I1203 13:33:07.493252 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerStarted","Data":"31ac604b3b9973bd8e383dc69aaff6758e2692c5d90e7d09a2b2aa1a2b830605"} Dec 03 13:33:07 crc kubenswrapper[4690]: I1203 13:33:07.836758 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.524087 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-696775c5d7-q6gwf"] Dec 03 13:33:09 crc kubenswrapper[4690]: E1203 13:33:09.533135 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerName="init" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.533180 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerName="init" Dec 03 13:33:09 crc kubenswrapper[4690]: E1203 13:33:09.533203 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerName="dnsmasq-dns" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.533211 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerName="dnsmasq-dns" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.533709 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8fd9d7d-4064-48c5-9a0a-cabf1de25647" containerName="dnsmasq-dns" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.555357 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.556052 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-696775c5d7-q6gwf"] Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.565962 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.566600 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.566719 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.597673 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9975bb19-d67e-4a25-a69f-8c1e37653961-etc-swift\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.597770 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9975bb19-d67e-4a25-a69f-8c1e37653961-log-httpd\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.597814 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9975bb19-d67e-4a25-a69f-8c1e37653961-run-httpd\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.597925 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngf4c\" (UniqueName: \"kubernetes.io/projected/9975bb19-d67e-4a25-a69f-8c1e37653961-kube-api-access-ngf4c\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.597971 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-combined-ca-bundle\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.598004 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-public-tls-certs\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.598070 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-internal-tls-certs\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.598205 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-config-data\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.699688 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngf4c\" (UniqueName: \"kubernetes.io/projected/9975bb19-d67e-4a25-a69f-8c1e37653961-kube-api-access-ngf4c\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.699751 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-combined-ca-bundle\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.699779 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-public-tls-certs\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.699818 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-internal-tls-certs\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.699904 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-config-data\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.699974 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9975bb19-d67e-4a25-a69f-8c1e37653961-etc-swift\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.700015 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9975bb19-d67e-4a25-a69f-8c1e37653961-log-httpd\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.700045 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9975bb19-d67e-4a25-a69f-8c1e37653961-run-httpd\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.700619 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9975bb19-d67e-4a25-a69f-8c1e37653961-run-httpd\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.700934 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9975bb19-d67e-4a25-a69f-8c1e37653961-log-httpd\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.706539 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-combined-ca-bundle\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.706678 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-internal-tls-certs\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.707629 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-config-data\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.708193 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9975bb19-d67e-4a25-a69f-8c1e37653961-public-tls-certs\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.716715 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngf4c\" (UniqueName: \"kubernetes.io/projected/9975bb19-d67e-4a25-a69f-8c1e37653961-kube-api-access-ngf4c\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.716976 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9975bb19-d67e-4a25-a69f-8c1e37653961-etc-swift\") pod \"swift-proxy-696775c5d7-q6gwf\" (UID: \"9975bb19-d67e-4a25-a69f-8c1e37653961\") " pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:09 crc kubenswrapper[4690]: I1203 13:33:09.892426 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:12 crc kubenswrapper[4690]: I1203 13:33:12.619068 4690 generic.go:334] "Generic (PLEG): container finished" podID="88d2e02a-af82-49f0-858a-9e9fa3dc4985" containerID="c052f9f5ff262734453b7188da832580e8edd7fd7095639527e4e1ba04504bed" exitCode=0 Dec 03 13:33:12 crc kubenswrapper[4690]: I1203 13:33:12.619164 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2rlzz" event={"ID":"88d2e02a-af82-49f0-858a-9e9fa3dc4985","Type":"ContainerDied","Data":"c052f9f5ff262734453b7188da832580e8edd7fd7095639527e4e1ba04504bed"} Dec 03 13:33:14 crc kubenswrapper[4690]: I1203 13:33:14.169764 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d8d5ff786-rbh82" Dec 03 13:33:14 crc kubenswrapper[4690]: I1203 13:33:14.281041 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8bf94c858-npzt6"] Dec 03 13:33:14 crc kubenswrapper[4690]: I1203 13:33:14.281595 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8bf94c858-npzt6" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api-log" containerID="cri-o://1ca6d398d142cc13c7aa5bb45a4f7f19ed37784935f0174787815fd5f45b0ebe" gracePeriod=30 Dec 03 13:33:14 crc kubenswrapper[4690]: I1203 13:33:14.281827 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8bf94c858-npzt6" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api" containerID="cri-o://f164137970f4783e844b6f624ab3958fdf03cf267f74f31ce3f5db4928eb0a9c" gracePeriod=30 Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.565631 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.644597 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-combined-ca-bundle\") pod \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.646844 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88d2e02a-af82-49f0-858a-9e9fa3dc4985-etc-machine-id\") pod \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.646917 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-config-data\") pod \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.647167 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-scripts\") pod \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.647235 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk8ps\" (UniqueName: \"kubernetes.io/projected/88d2e02a-af82-49f0-858a-9e9fa3dc4985-kube-api-access-sk8ps\") pod \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.647325 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-db-sync-config-data\") pod \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\" (UID: \"88d2e02a-af82-49f0-858a-9e9fa3dc4985\") " Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.655307 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88d2e02a-af82-49f0-858a-9e9fa3dc4985-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "88d2e02a-af82-49f0-858a-9e9fa3dc4985" (UID: "88d2e02a-af82-49f0-858a-9e9fa3dc4985"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.659286 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "88d2e02a-af82-49f0-858a-9e9fa3dc4985" (UID: "88d2e02a-af82-49f0-858a-9e9fa3dc4985"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.665179 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88d2e02a-af82-49f0-858a-9e9fa3dc4985-kube-api-access-sk8ps" (OuterVolumeSpecName: "kube-api-access-sk8ps") pod "88d2e02a-af82-49f0-858a-9e9fa3dc4985" (UID: "88d2e02a-af82-49f0-858a-9e9fa3dc4985"). InnerVolumeSpecName "kube-api-access-sk8ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.679461 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2rlzz" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.679452 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2rlzz" event={"ID":"88d2e02a-af82-49f0-858a-9e9fa3dc4985","Type":"ContainerDied","Data":"a82ed63ec10f55dcee76d3065d231f9b07e90270c8ab3d45613f47efc9e874f1"} Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.679622 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a82ed63ec10f55dcee76d3065d231f9b07e90270c8ab3d45613f47efc9e874f1" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.680132 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-scripts" (OuterVolumeSpecName: "scripts") pod "88d2e02a-af82-49f0-858a-9e9fa3dc4985" (UID: "88d2e02a-af82-49f0-858a-9e9fa3dc4985"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.682163 4690 generic.go:334] "Generic (PLEG): container finished" podID="a683ac29-783f-4210-97dc-42b0deff462d" containerID="1ca6d398d142cc13c7aa5bb45a4f7f19ed37784935f0174787815fd5f45b0ebe" exitCode=143 Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.682215 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8bf94c858-npzt6" event={"ID":"a683ac29-783f-4210-97dc-42b0deff462d","Type":"ContainerDied","Data":"1ca6d398d142cc13c7aa5bb45a4f7f19ed37784935f0174787815fd5f45b0ebe"} Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.713279 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88d2e02a-af82-49f0-858a-9e9fa3dc4985" (UID: "88d2e02a-af82-49f0-858a-9e9fa3dc4985"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.740596 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-config-data" (OuterVolumeSpecName: "config-data") pod "88d2e02a-af82-49f0-858a-9e9fa3dc4985" (UID: "88d2e02a-af82-49f0-858a-9e9fa3dc4985"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.750189 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.750458 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk8ps\" (UniqueName: \"kubernetes.io/projected/88d2e02a-af82-49f0-858a-9e9fa3dc4985-kube-api-access-sk8ps\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.750581 4690 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.750668 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.750731 4690 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88d2e02a-af82-49f0-858a-9e9fa3dc4985-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.750813 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d2e02a-af82-49f0-858a-9e9fa3dc4985-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:15 crc kubenswrapper[4690]: I1203 13:33:15.966161 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-696775c5d7-q6gwf"] Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.709102 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerStarted","Data":"6ae82468b738e6a64295012886490c63cc488a690beeb949309d80e27b75f266"} Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.711811 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f47b845f-82f3-44b9-966d-eb3076f05362","Type":"ContainerStarted","Data":"d7535c5da86c8f4bc36e4089313ad65a4dee4b7854d0316a09ff83fe9dd5f130"} Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.716766 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696775c5d7-q6gwf" event={"ID":"9975bb19-d67e-4a25-a69f-8c1e37653961","Type":"ContainerStarted","Data":"57652ccf30fae5cc9bb3a2aa0d3ad7ded4823a3710511513f6597db991fc66c0"} Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.716827 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696775c5d7-q6gwf" event={"ID":"9975bb19-d67e-4a25-a69f-8c1e37653961","Type":"ContainerStarted","Data":"ba5d574e88c6f4dde780ed409503a079df8503550e5971f7901b9a049357b23b"} Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.716842 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696775c5d7-q6gwf" event={"ID":"9975bb19-d67e-4a25-a69f-8c1e37653961","Type":"ContainerStarted","Data":"56a0705b630f8984dcf1457a35ca03bfd24b9a23135b09aaeee077792a404904"} Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.717277 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.717814 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.733591 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.734046 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-httpd" containerID="cri-o://d030d61740116392c7f29accd15b1953269757d2f02a2a14d263e9faf60b1675" gracePeriod=30 Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.734226 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-log" containerID="cri-o://6146c38d3aacfc70340716008a2714e4faa89e7d667db18a56c938df48dbf6d0" gracePeriod=30 Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.743463 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.622425626 podStartE2EDuration="19.743439956s" podCreationTimestamp="2025-12-03 13:32:57 +0000 UTC" firstStartedPulling="2025-12-03 13:32:59.488232433 +0000 UTC m=+1425.469152866" lastFinishedPulling="2025-12-03 13:33:15.609246773 +0000 UTC m=+1441.590167196" observedRunningTime="2025-12-03 13:33:16.740030027 +0000 UTC m=+1442.720950470" watchObservedRunningTime="2025-12-03 13:33:16.743439956 +0000 UTC m=+1442.724360389" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.771136 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-696775c5d7-q6gwf" podStartSLOduration=7.771118612 podStartE2EDuration="7.771118612s" podCreationTimestamp="2025-12-03 13:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:16.769914391 +0000 UTC m=+1442.750834834" watchObservedRunningTime="2025-12-03 13:33:16.771118612 +0000 UTC m=+1442.752039045" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.823956 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.824020 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.881492 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:16 crc kubenswrapper[4690]: E1203 13:33:16.882934 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d2e02a-af82-49f0-858a-9e9fa3dc4985" containerName="cinder-db-sync" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.882957 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d2e02a-af82-49f0-858a-9e9fa3dc4985" containerName="cinder-db-sync" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.883222 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="88d2e02a-af82-49f0-858a-9e9fa3dc4985" containerName="cinder-db-sync" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.884397 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.887121 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-lqrtb" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.887461 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.887650 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.887803 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.900187 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.953945 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-gm28d"] Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.956024 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987301 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987382 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987443 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-config\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987498 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe83b063-4c0b-4a16-8460-da53de551d17-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987604 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987639 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st47z\" (UniqueName: \"kubernetes.io/projected/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-kube-api-access-st47z\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987691 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987720 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.987825 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.988031 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.988078 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k55tb\" (UniqueName: \"kubernetes.io/projected/fe83b063-4c0b-4a16-8460-da53de551d17-kube-api-access-k55tb\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.988149 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:16 crc kubenswrapper[4690]: I1203 13:33:16.997462 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-gm28d"] Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.072638 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.078535 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.096701 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.098611 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.100748 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.100829 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st47z\" (UniqueName: \"kubernetes.io/projected/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-kube-api-access-st47z\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.100878 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.100961 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.100991 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101041 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24c086e4-0b3c-423d-ab72-7ca515d60cc4-logs\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101136 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101185 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24c086e4-0b3c-423d-ab72-7ca515d60cc4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101212 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101265 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101289 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data-custom\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101331 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k55tb\" (UniqueName: \"kubernetes.io/projected/fe83b063-4c0b-4a16-8460-da53de551d17-kube-api-access-k55tb\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101371 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-scripts\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101435 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101527 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101562 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101590 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-config\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101642 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe83b063-4c0b-4a16-8460-da53de551d17-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.101663 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b4zj\" (UniqueName: \"kubernetes.io/projected/24c086e4-0b3c-423d-ab72-7ca515d60cc4-kube-api-access-9b4zj\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.109629 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.111909 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe83b063-4c0b-4a16-8460-da53de551d17-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.112711 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-config\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.112900 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.116734 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.120806 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.121491 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.128407 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.130599 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.135024 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.137807 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st47z\" (UniqueName: \"kubernetes.io/projected/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-kube-api-access-st47z\") pod \"dnsmasq-dns-795f4db4bc-gm28d\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.145230 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k55tb\" (UniqueName: \"kubernetes.io/projected/fe83b063-4c0b-4a16-8460-da53de551d17-kube-api-access-k55tb\") pod \"cinder-scheduler-0\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.205172 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24c086e4-0b3c-423d-ab72-7ca515d60cc4-logs\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.205272 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24c086e4-0b3c-423d-ab72-7ca515d60cc4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.205294 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.205329 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data-custom\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.205356 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-scripts\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.205439 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b4zj\" (UniqueName: \"kubernetes.io/projected/24c086e4-0b3c-423d-ab72-7ca515d60cc4-kube-api-access-9b4zj\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.205501 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.208891 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24c086e4-0b3c-423d-ab72-7ca515d60cc4-logs\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.209094 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24c086e4-0b3c-423d-ab72-7ca515d60cc4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.210209 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.218536 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data-custom\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.225603 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-scripts\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.232232 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.243565 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b4zj\" (UniqueName: \"kubernetes.io/projected/24c086e4-0b3c-423d-ab72-7ca515d60cc4-kube-api-access-9b4zj\") pod \"cinder-api-0\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.269613 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.429376 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.451432 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.491049 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8bf94c858-npzt6" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:58346->10.217.0.166:9311: read: connection reset by peer" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.491438 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8bf94c858-npzt6" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:58344->10.217.0.166:9311: read: connection reset by peer" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.747699 4690 generic.go:334] "Generic (PLEG): container finished" podID="a683ac29-783f-4210-97dc-42b0deff462d" containerID="f164137970f4783e844b6f624ab3958fdf03cf267f74f31ce3f5db4928eb0a9c" exitCode=0 Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.747792 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8bf94c858-npzt6" event={"ID":"a683ac29-783f-4210-97dc-42b0deff462d","Type":"ContainerDied","Data":"f164137970f4783e844b6f624ab3958fdf03cf267f74f31ce3f5db4928eb0a9c"} Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.759247 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerStarted","Data":"9c6f89be5976ed0d34f8ab2aa5e740da117e870e128be0d0bab58fa754224aff"} Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.760292 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.779495 4690 generic.go:334] "Generic (PLEG): container finished" podID="604fdece-2b3e-4786-995d-68d784e5d263" containerID="6146c38d3aacfc70340716008a2714e4faa89e7d667db18a56c938df48dbf6d0" exitCode=143 Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.779612 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"604fdece-2b3e-4786-995d-68d784e5d263","Type":"ContainerDied","Data":"6146c38d3aacfc70340716008a2714e4faa89e7d667db18a56c938df48dbf6d0"} Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.806509 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.413416567 podStartE2EDuration="22.806483138s" podCreationTimestamp="2025-12-03 13:32:55 +0000 UTC" firstStartedPulling="2025-12-03 13:32:59.576419035 +0000 UTC m=+1425.557339458" lastFinishedPulling="2025-12-03 13:33:16.969485596 +0000 UTC m=+1442.950406029" observedRunningTime="2025-12-03 13:33:17.78336853 +0000 UTC m=+1443.764288983" watchObservedRunningTime="2025-12-03 13:33:17.806483138 +0000 UTC m=+1443.787403571" Dec 03 13:33:17 crc kubenswrapper[4690]: I1203 13:33:17.954454 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.160612 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.395020 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.421591 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-gm28d"] Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.526431 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data-custom\") pod \"a683ac29-783f-4210-97dc-42b0deff462d\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.526591 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9m6k\" (UniqueName: \"kubernetes.io/projected/a683ac29-783f-4210-97dc-42b0deff462d-kube-api-access-c9m6k\") pod \"a683ac29-783f-4210-97dc-42b0deff462d\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.526674 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data\") pod \"a683ac29-783f-4210-97dc-42b0deff462d\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.526762 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a683ac29-783f-4210-97dc-42b0deff462d-logs\") pod \"a683ac29-783f-4210-97dc-42b0deff462d\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.527063 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-combined-ca-bundle\") pod \"a683ac29-783f-4210-97dc-42b0deff462d\" (UID: \"a683ac29-783f-4210-97dc-42b0deff462d\") " Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.573326 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a683ac29-783f-4210-97dc-42b0deff462d" (UID: "a683ac29-783f-4210-97dc-42b0deff462d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.573687 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a683ac29-783f-4210-97dc-42b0deff462d-kube-api-access-c9m6k" (OuterVolumeSpecName: "kube-api-access-c9m6k") pod "a683ac29-783f-4210-97dc-42b0deff462d" (UID: "a683ac29-783f-4210-97dc-42b0deff462d"). InnerVolumeSpecName "kube-api-access-c9m6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.586928 4690 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.586970 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9m6k\" (UniqueName: \"kubernetes.io/projected/a683ac29-783f-4210-97dc-42b0deff462d-kube-api-access-c9m6k\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.588887 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a683ac29-783f-4210-97dc-42b0deff462d-logs" (OuterVolumeSpecName: "logs") pod "a683ac29-783f-4210-97dc-42b0deff462d" (UID: "a683ac29-783f-4210-97dc-42b0deff462d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.616547 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a683ac29-783f-4210-97dc-42b0deff462d" (UID: "a683ac29-783f-4210-97dc-42b0deff462d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.696358 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.696388 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a683ac29-783f-4210-97dc-42b0deff462d-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.777030 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data" (OuterVolumeSpecName: "config-data") pod "a683ac29-783f-4210-97dc-42b0deff462d" (UID: "a683ac29-783f-4210-97dc-42b0deff462d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.799712 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a683ac29-783f-4210-97dc-42b0deff462d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.817008 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" event={"ID":"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b","Type":"ContainerStarted","Data":"84393b150ef0c8b61ef0aa3183f57559203ffceac415ea6c2313aece80bec53f"} Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.819595 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8bf94c858-npzt6" event={"ID":"a683ac29-783f-4210-97dc-42b0deff462d","Type":"ContainerDied","Data":"48ac755eb751d3ffd95c6ab82d52100f2002246baa335c37305d8a98023b8608"} Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.819641 4690 scope.go:117] "RemoveContainer" containerID="f164137970f4783e844b6f624ab3958fdf03cf267f74f31ce3f5db4928eb0a9c" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.819811 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8bf94c858-npzt6" Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.830839 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"24c086e4-0b3c-423d-ab72-7ca515d60cc4","Type":"ContainerStarted","Data":"43f4ec465ee0bb0a53593cf05ea975e672d598997fd4927c5c8c68bc224a6833"} Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.837083 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe83b063-4c0b-4a16-8460-da53de551d17","Type":"ContainerStarted","Data":"83e8f690e6cbd43bfd04b810eb0beba9d22b0d73408e08032425ef03dd17b51e"} Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.878555 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8bf94c858-npzt6"] Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.893042 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-8bf94c858-npzt6"] Dec 03 13:33:18 crc kubenswrapper[4690]: I1203 13:33:18.902566 4690 scope.go:117] "RemoveContainer" containerID="1ca6d398d142cc13c7aa5bb45a4f7f19ed37784935f0174787815fd5f45b0ebe" Dec 03 13:33:19 crc kubenswrapper[4690]: I1203 13:33:19.299209 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:20 crc kubenswrapper[4690]: W1203 13:33:20.253886 4690 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4750fc4_e086_4be9_a5e8_aa7c0f778f6b.slice/crio-c7258d23bd6f397b55c6a1061f67a8ed53fe23e846e619f5e095335cd053ffd6.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4750fc4_e086_4be9_a5e8_aa7c0f778f6b.slice/crio-c7258d23bd6f397b55c6a1061f67a8ed53fe23e846e619f5e095335cd053ffd6.scope: no such file or directory Dec 03 13:33:20 crc kubenswrapper[4690]: I1203 13:33:20.336944 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a683ac29-783f-4210-97dc-42b0deff462d" path="/var/lib/kubelet/pods/a683ac29-783f-4210-97dc-42b0deff462d/volumes" Dec 03 13:33:20 crc kubenswrapper[4690]: E1203 13:33:20.535421 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88d2e02a_af82_49f0_858a_9e9fa3dc4985.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda683ac29_783f_4210_97dc_42b0deff462d.slice/crio-f164137970f4783e844b6f624ab3958fdf03cf267f74f31ce3f5db4928eb0a9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604fdece_2b3e_4786_995d_68d784e5d263.slice/crio-conmon-6146c38d3aacfc70340716008a2714e4faa89e7d667db18a56c938df48dbf6d0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda683ac29_783f_4210_97dc_42b0deff462d.slice/crio-48ac755eb751d3ffd95c6ab82d52100f2002246baa335c37305d8a98023b8608\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88d2e02a_af82_49f0_858a_9e9fa3dc4985.slice/crio-a82ed63ec10f55dcee76d3065d231f9b07e90270c8ab3d45613f47efc9e874f1\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604fdece_2b3e_4786_995d_68d784e5d263.slice/crio-d030d61740116392c7f29accd15b1953269757d2f02a2a14d263e9faf60b1675.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604fdece_2b3e_4786_995d_68d784e5d263.slice/crio-6146c38d3aacfc70340716008a2714e4faa89e7d667db18a56c938df48dbf6d0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda683ac29_783f_4210_97dc_42b0deff462d.slice/crio-conmon-f164137970f4783e844b6f624ab3958fdf03cf267f74f31ce3f5db4928eb0a9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda683ac29_783f_4210_97dc_42b0deff462d.slice\": RecentStats: unable to find data in memory cache]" Dec 03 13:33:20 crc kubenswrapper[4690]: I1203 13:33:20.869822 4690 generic.go:334] "Generic (PLEG): container finished" podID="604fdece-2b3e-4786-995d-68d784e5d263" containerID="d030d61740116392c7f29accd15b1953269757d2f02a2a14d263e9faf60b1675" exitCode=0 Dec 03 13:33:20 crc kubenswrapper[4690]: I1203 13:33:20.869934 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"604fdece-2b3e-4786-995d-68d784e5d263","Type":"ContainerDied","Data":"d030d61740116392c7f29accd15b1953269757d2f02a2a14d263e9faf60b1675"} Dec 03 13:33:20 crc kubenswrapper[4690]: I1203 13:33:20.872107 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"24c086e4-0b3c-423d-ab72-7ca515d60cc4","Type":"ContainerStarted","Data":"f60c2363d6c12bbc778debe0553e86e82352c5271d97bb311d8e223dd17d5f2f"} Dec 03 13:33:20 crc kubenswrapper[4690]: I1203 13:33:20.874061 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" containerID="c7258d23bd6f397b55c6a1061f67a8ed53fe23e846e619f5e095335cd053ffd6" exitCode=0 Dec 03 13:33:20 crc kubenswrapper[4690]: I1203 13:33:20.874109 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" event={"ID":"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b","Type":"ContainerDied","Data":"c7258d23bd6f397b55c6a1061f67a8ed53fe23e846e619f5e095335cd053ffd6"} Dec 03 13:33:20 crc kubenswrapper[4690]: I1203 13:33:20.929353 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.039147 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.039442 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerName="glance-log" containerID="cri-o://49017db7e96374547b2aab24e42d9d98a9647f40c8f34a5ef74506765384840c" gracePeriod=30 Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.040003 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerName="glance-httpd" containerID="cri-o://50fb8514b26b86d905c80d505bfd3f7977b619a006d3c4bd64746853de8bcb53" gracePeriod=30 Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.048440 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7f775c89b8-4tcrc" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.229112 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.160:9292/healthcheck\": dial tcp 10.217.0.160:9292: connect: connection refused" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.230705 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.160:9292/healthcheck\": dial tcp 10.217.0.160:9292: connect: connection refused" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.834182 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.916970 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"604fdece-2b3e-4786-995d-68d784e5d263\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.917326 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-logs\") pod \"604fdece-2b3e-4786-995d-68d784e5d263\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.917524 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-httpd-run\") pod \"604fdece-2b3e-4786-995d-68d784e5d263\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.917589 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-scripts\") pod \"604fdece-2b3e-4786-995d-68d784e5d263\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.917629 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwmd6\" (UniqueName: \"kubernetes.io/projected/604fdece-2b3e-4786-995d-68d784e5d263-kube-api-access-pwmd6\") pod \"604fdece-2b3e-4786-995d-68d784e5d263\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.917685 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-combined-ca-bundle\") pod \"604fdece-2b3e-4786-995d-68d784e5d263\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.917725 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-config-data\") pod \"604fdece-2b3e-4786-995d-68d784e5d263\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.917747 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-public-tls-certs\") pod \"604fdece-2b3e-4786-995d-68d784e5d263\" (UID: \"604fdece-2b3e-4786-995d-68d784e5d263\") " Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.924680 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-logs" (OuterVolumeSpecName: "logs") pod "604fdece-2b3e-4786-995d-68d784e5d263" (UID: "604fdece-2b3e-4786-995d-68d784e5d263"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.926415 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/604fdece-2b3e-4786-995d-68d784e5d263-kube-api-access-pwmd6" (OuterVolumeSpecName: "kube-api-access-pwmd6") pod "604fdece-2b3e-4786-995d-68d784e5d263" (UID: "604fdece-2b3e-4786-995d-68d784e5d263"). InnerVolumeSpecName "kube-api-access-pwmd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.926708 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "604fdece-2b3e-4786-995d-68d784e5d263" (UID: "604fdece-2b3e-4786-995d-68d784e5d263"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.926901 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"24c086e4-0b3c-423d-ab72-7ca515d60cc4","Type":"ContainerStarted","Data":"023b76b379606e17230fe58d269083fca552cf24af0168971af0d9f5f87ec99a"} Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.927070 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerName="cinder-api-log" containerID="cri-o://f60c2363d6c12bbc778debe0553e86e82352c5271d97bb311d8e223dd17d5f2f" gracePeriod=30 Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.927148 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.927502 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerName="cinder-api" containerID="cri-o://023b76b379606e17230fe58d269083fca552cf24af0168971af0d9f5f87ec99a" gracePeriod=30 Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.930527 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "604fdece-2b3e-4786-995d-68d784e5d263" (UID: "604fdece-2b3e-4786-995d-68d784e5d263"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.930626 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-scripts" (OuterVolumeSpecName: "scripts") pod "604fdece-2b3e-4786-995d-68d784e5d263" (UID: "604fdece-2b3e-4786-995d-68d784e5d263"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.971408 4690 generic.go:334] "Generic (PLEG): container finished" podID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerID="49017db7e96374547b2aab24e42d9d98a9647f40c8f34a5ef74506765384840c" exitCode=143 Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.971617 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"105e22f6-36c9-4ab3-8e64-9f056a266704","Type":"ContainerDied","Data":"49017db7e96374547b2aab24e42d9d98a9647f40c8f34a5ef74506765384840c"} Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.981026 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.980976135 podStartE2EDuration="4.980976135s" podCreationTimestamp="2025-12-03 13:33:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:21.966201193 +0000 UTC m=+1447.947121636" watchObservedRunningTime="2025-12-03 13:33:21.980976135 +0000 UTC m=+1447.961896568" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.982485 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"604fdece-2b3e-4786-995d-68d784e5d263","Type":"ContainerDied","Data":"9d0d9e980146105b0e63b7a4cb5e155396d85ab2d3eaa573643f6d48db602fb4"} Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.982553 4690 scope.go:117] "RemoveContainer" containerID="d030d61740116392c7f29accd15b1953269757d2f02a2a14d263e9faf60b1675" Dec 03 13:33:21 crc kubenswrapper[4690]: I1203 13:33:21.983066 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.001160 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "604fdece-2b3e-4786-995d-68d784e5d263" (UID: "604fdece-2b3e-4786-995d-68d784e5d263"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.010528 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "604fdece-2b3e-4786-995d-68d784e5d263" (UID: "604fdece-2b3e-4786-995d-68d784e5d263"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.021610 4690 scope.go:117] "RemoveContainer" containerID="6146c38d3aacfc70340716008a2714e4faa89e7d667db18a56c938df48dbf6d0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.025336 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.025391 4690 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.025402 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.025489 4690 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.025502 4690 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/604fdece-2b3e-4786-995d-68d784e5d263-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.025530 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.025539 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwmd6\" (UniqueName: \"kubernetes.io/projected/604fdece-2b3e-4786-995d-68d784e5d263-kube-api-access-pwmd6\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.057284 4690 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.069281 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-config-data" (OuterVolumeSpecName: "config-data") pod "604fdece-2b3e-4786-995d-68d784e5d263" (UID: "604fdece-2b3e-4786-995d-68d784e5d263"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.127389 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604fdece-2b3e-4786-995d-68d784e5d263-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.127427 4690 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.330716 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.347201 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.371586 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:33:22 crc kubenswrapper[4690]: E1203 13:33:22.383188 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-httpd" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.383217 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-httpd" Dec 03 13:33:22 crc kubenswrapper[4690]: E1203 13:33:22.383242 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api-log" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.383249 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api-log" Dec 03 13:33:22 crc kubenswrapper[4690]: E1203 13:33:22.383258 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-log" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.383264 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-log" Dec 03 13:33:22 crc kubenswrapper[4690]: E1203 13:33:22.383281 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.383290 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.383653 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-httpd" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.383671 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.383695 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a683ac29-783f-4210-97dc-42b0deff462d" containerName="barbican-api-log" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.383706 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="604fdece-2b3e-4786-995d-68d784e5d263" containerName="glance-log" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.384951 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.389845 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.391145 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.428395 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.431687 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8t62\" (UniqueName: \"kubernetes.io/projected/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-kube-api-access-m8t62\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.431737 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.431776 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.431825 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.431877 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-logs\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.431917 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-scripts\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.431967 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.432022 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-config-data\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533160 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-config-data\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533227 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8t62\" (UniqueName: \"kubernetes.io/projected/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-kube-api-access-m8t62\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533262 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533294 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533343 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533391 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-logs\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533433 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-scripts\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533461 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.533885 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.535459 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.535684 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-logs\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.539939 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-scripts\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.540740 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.541314 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-config-data\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.541924 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.556950 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8t62\" (UniqueName: \"kubernetes.io/projected/ada5ac2f-e8b5-4437-b639-183fe2bf3c09-kube-api-access-m8t62\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.567269 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ada5ac2f-e8b5-4437-b639-183fe2bf3c09\") " pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.725724 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.994373 4690 generic.go:334] "Generic (PLEG): container finished" podID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerID="023b76b379606e17230fe58d269083fca552cf24af0168971af0d9f5f87ec99a" exitCode=0 Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.994745 4690 generic.go:334] "Generic (PLEG): container finished" podID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerID="f60c2363d6c12bbc778debe0553e86e82352c5271d97bb311d8e223dd17d5f2f" exitCode=143 Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.994456 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"24c086e4-0b3c-423d-ab72-7ca515d60cc4","Type":"ContainerDied","Data":"023b76b379606e17230fe58d269083fca552cf24af0168971af0d9f5f87ec99a"} Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.994828 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"24c086e4-0b3c-423d-ab72-7ca515d60cc4","Type":"ContainerDied","Data":"f60c2363d6c12bbc778debe0553e86e82352c5271d97bb311d8e223dd17d5f2f"} Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.997220 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c7c4b9c77-f54ln" event={"ID":"f7a19182-b617-47e3-bc7f-db372257b984","Type":"ContainerStarted","Data":"fcb7b848595027754cbac6acb34c1a0b1fa9269b9c5489abd341a1591c6caa53"} Dec 03 13:33:22 crc kubenswrapper[4690]: I1203 13:33:22.997248 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c7c4b9c77-f54ln" event={"ID":"f7a19182-b617-47e3-bc7f-db372257b984","Type":"ContainerStarted","Data":"fe3d841412dffea7f263f3e1fbb870e4cb70c2df7cd6aa70ed59ec5ddba8e507"} Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:22.999500 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe83b063-4c0b-4a16-8460-da53de551d17","Type":"ContainerStarted","Data":"4ba06fc7351cb8d2092580e2d1d94aa907ad00d987372a1856d65b3476db9555"} Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.003603 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" event={"ID":"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b","Type":"ContainerStarted","Data":"6d0f45e48fb229e3a9ff169b79841cc3548277162151de824334818215f6cb23"} Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.003733 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.028278 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7c7c4b9c77-f54ln" podStartSLOduration=3.010855778 podStartE2EDuration="33.02824826s" podCreationTimestamp="2025-12-03 13:32:50 +0000 UTC" firstStartedPulling="2025-12-03 13:32:51.431091492 +0000 UTC m=+1417.412011925" lastFinishedPulling="2025-12-03 13:33:21.448483974 +0000 UTC m=+1447.429404407" observedRunningTime="2025-12-03 13:33:23.011896296 +0000 UTC m=+1448.992816739" watchObservedRunningTime="2025-12-03 13:33:23.02824826 +0000 UTC m=+1449.009168693" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.063029 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" podStartSLOduration=7.063001979 podStartE2EDuration="7.063001979s" podCreationTimestamp="2025-12-03 13:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:23.048689629 +0000 UTC m=+1449.029610072" watchObservedRunningTime="2025-12-03 13:33:23.063001979 +0000 UTC m=+1449.043922412" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.301999 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.666716 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.760037 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-scripts\") pod \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.760135 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24c086e4-0b3c-423d-ab72-7ca515d60cc4-logs\") pod \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.760182 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24c086e4-0b3c-423d-ab72-7ca515d60cc4-etc-machine-id\") pod \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.760219 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data-custom\") pod \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.760388 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b4zj\" (UniqueName: \"kubernetes.io/projected/24c086e4-0b3c-423d-ab72-7ca515d60cc4-kube-api-access-9b4zj\") pod \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.760422 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-combined-ca-bundle\") pod \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.760442 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data\") pod \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\" (UID: \"24c086e4-0b3c-423d-ab72-7ca515d60cc4\") " Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.761064 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24c086e4-0b3c-423d-ab72-7ca515d60cc4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "24c086e4-0b3c-423d-ab72-7ca515d60cc4" (UID: "24c086e4-0b3c-423d-ab72-7ca515d60cc4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.761257 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24c086e4-0b3c-423d-ab72-7ca515d60cc4-logs" (OuterVolumeSpecName: "logs") pod "24c086e4-0b3c-423d-ab72-7ca515d60cc4" (UID: "24c086e4-0b3c-423d-ab72-7ca515d60cc4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.765103 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24c086e4-0b3c-423d-ab72-7ca515d60cc4-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.765130 4690 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24c086e4-0b3c-423d-ab72-7ca515d60cc4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.784149 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "24c086e4-0b3c-423d-ab72-7ca515d60cc4" (UID: "24c086e4-0b3c-423d-ab72-7ca515d60cc4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.784240 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24c086e4-0b3c-423d-ab72-7ca515d60cc4-kube-api-access-9b4zj" (OuterVolumeSpecName: "kube-api-access-9b4zj") pod "24c086e4-0b3c-423d-ab72-7ca515d60cc4" (UID: "24c086e4-0b3c-423d-ab72-7ca515d60cc4"). InnerVolumeSpecName "kube-api-access-9b4zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.794084 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-scripts" (OuterVolumeSpecName: "scripts") pod "24c086e4-0b3c-423d-ab72-7ca515d60cc4" (UID: "24c086e4-0b3c-423d-ab72-7ca515d60cc4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.807733 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24c086e4-0b3c-423d-ab72-7ca515d60cc4" (UID: "24c086e4-0b3c-423d-ab72-7ca515d60cc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.834209 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data" (OuterVolumeSpecName: "config-data") pod "24c086e4-0b3c-423d-ab72-7ca515d60cc4" (UID: "24c086e4-0b3c-423d-ab72-7ca515d60cc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.866676 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b4zj\" (UniqueName: \"kubernetes.io/projected/24c086e4-0b3c-423d-ab72-7ca515d60cc4-kube-api-access-9b4zj\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.866711 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.866723 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.866732 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:23 crc kubenswrapper[4690]: I1203 13:33:23.866743 4690 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24c086e4-0b3c-423d-ab72-7ca515d60cc4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.019279 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"24c086e4-0b3c-423d-ab72-7ca515d60cc4","Type":"ContainerDied","Data":"43f4ec465ee0bb0a53593cf05ea975e672d598997fd4927c5c8c68bc224a6833"} Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.019334 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.019719 4690 scope.go:117] "RemoveContainer" containerID="023b76b379606e17230fe58d269083fca552cf24af0168971af0d9f5f87ec99a" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.021267 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ada5ac2f-e8b5-4437-b639-183fe2bf3c09","Type":"ContainerStarted","Data":"0b5fff42ca485e44116c4e1e5c51f9cc0bc155bfd9f11fdd7a83811510ec34b0"} Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.024791 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe83b063-4c0b-4a16-8460-da53de551d17","Type":"ContainerStarted","Data":"0f8d67945b6f649416d68ba70fffbb38afe567ab4b4c2d82d166e064f69a7b04"} Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.065889 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.088797 4690 scope.go:117] "RemoveContainer" containerID="f60c2363d6c12bbc778debe0553e86e82352c5271d97bb311d8e223dd17d5f2f" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.112006 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.120667 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:24 crc kubenswrapper[4690]: E1203 13:33:24.121288 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerName="cinder-api-log" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.121305 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerName="cinder-api-log" Dec 03 13:33:24 crc kubenswrapper[4690]: E1203 13:33:24.121370 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerName="cinder-api" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.121381 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerName="cinder-api" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.121611 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerName="cinder-api-log" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.121623 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" containerName="cinder-api" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.123048 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.126314 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.126514 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.126668 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.129578 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.185616 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-config-data-custom\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.185699 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-config-data\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.185790 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d33ca8c1-d71d-4624-a362-f845fda14976-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.185880 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.185958 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mflxq\" (UniqueName: \"kubernetes.io/projected/d33ca8c1-d71d-4624-a362-f845fda14976-kube-api-access-mflxq\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.185999 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.186020 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-scripts\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.186035 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.186091 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33ca8c1-d71d-4624-a362-f845fda14976-logs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.288261 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.288318 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-scripts\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.288339 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.288423 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33ca8c1-d71d-4624-a362-f845fda14976-logs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.288442 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-config-data-custom\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.288479 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-config-data\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.288936 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33ca8c1-d71d-4624-a362-f845fda14976-logs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.289141 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d33ca8c1-d71d-4624-a362-f845fda14976-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.289186 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.289235 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mflxq\" (UniqueName: \"kubernetes.io/projected/d33ca8c1-d71d-4624-a362-f845fda14976-kube-api-access-mflxq\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.289367 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d33ca8c1-d71d-4624-a362-f845fda14976-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.295364 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-config-data-custom\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.295612 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.299247 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.315877 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-scripts\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.318820 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-config-data\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.322425 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d33ca8c1-d71d-4624-a362-f845fda14976-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.322955 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mflxq\" (UniqueName: \"kubernetes.io/projected/d33ca8c1-d71d-4624-a362-f845fda14976-kube-api-access-mflxq\") pod \"cinder-api-0\" (UID: \"d33ca8c1-d71d-4624-a362-f845fda14976\") " pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.330603 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24c086e4-0b3c-423d-ab72-7ca515d60cc4" path="/var/lib/kubelet/pods/24c086e4-0b3c-423d-ab72-7ca515d60cc4/volumes" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.331318 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="604fdece-2b3e-4786-995d-68d784e5d263" path="/var/lib/kubelet/pods/604fdece-2b3e-4786-995d-68d784e5d263/volumes" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.447072 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.903298 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:24 crc kubenswrapper[4690]: I1203 13:33:24.904115 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-696775c5d7-q6gwf" Dec 03 13:33:25 crc kubenswrapper[4690]: I1203 13:33:25.001977 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:33:25 crc kubenswrapper[4690]: I1203 13:33:25.096841 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33ca8c1-d71d-4624-a362-f845fda14976","Type":"ContainerStarted","Data":"9c594ec862a198af4a6c1d8af8e11339762b5f92bf3ca9369292df8ce9fac3ed"} Dec 03 13:33:25 crc kubenswrapper[4690]: I1203 13:33:25.389835 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:25 crc kubenswrapper[4690]: I1203 13:33:25.390605 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="proxy-httpd" containerID="cri-o://9c6f89be5976ed0d34f8ab2aa5e740da117e870e128be0d0bab58fa754224aff" gracePeriod=30 Dec 03 13:33:25 crc kubenswrapper[4690]: I1203 13:33:25.390977 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="sg-core" containerID="cri-o://6ae82468b738e6a64295012886490c63cc488a690beeb949309d80e27b75f266" gracePeriod=30 Dec 03 13:33:25 crc kubenswrapper[4690]: I1203 13:33:25.391049 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="ceilometer-notification-agent" containerID="cri-o://31ac604b3b9973bd8e383dc69aaff6758e2692c5d90e7d09a2b2aa1a2b830605" gracePeriod=30 Dec 03 13:33:25 crc kubenswrapper[4690]: I1203 13:33:25.391134 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="ceilometer-central-agent" containerID="cri-o://294b0b57fa1be9e729837ce3dda9c0dff30aab373239cb0aacc2f8d55c41747c" gracePeriod=30 Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.166150 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ada5ac2f-e8b5-4437-b639-183fe2bf3c09","Type":"ContainerStarted","Data":"18f59089f21fd395f65f76dcd238131658bc596c45a9163ddd361e6d7e711465"} Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.171443 4690 generic.go:334] "Generic (PLEG): container finished" podID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerID="50fb8514b26b86d905c80d505bfd3f7977b619a006d3c4bd64746853de8bcb53" exitCode=0 Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.171517 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"105e22f6-36c9-4ab3-8e64-9f056a266704","Type":"ContainerDied","Data":"50fb8514b26b86d905c80d505bfd3f7977b619a006d3c4bd64746853de8bcb53"} Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.175276 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerID="9c6f89be5976ed0d34f8ab2aa5e740da117e870e128be0d0bab58fa754224aff" exitCode=0 Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.175294 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerID="6ae82468b738e6a64295012886490c63cc488a690beeb949309d80e27b75f266" exitCode=2 Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.175303 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerID="294b0b57fa1be9e729837ce3dda9c0dff30aab373239cb0aacc2f8d55c41747c" exitCode=0 Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.175308 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerDied","Data":"9c6f89be5976ed0d34f8ab2aa5e740da117e870e128be0d0bab58fa754224aff"} Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.175522 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerDied","Data":"6ae82468b738e6a64295012886490c63cc488a690beeb949309d80e27b75f266"} Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.175555 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerDied","Data":"294b0b57fa1be9e729837ce3dda9c0dff30aab373239cb0aacc2f8d55c41747c"} Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.191522 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.214280 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.748793997 podStartE2EDuration="10.214255584s" podCreationTimestamp="2025-12-03 13:33:16 +0000 UTC" firstStartedPulling="2025-12-03 13:33:17.968039879 +0000 UTC m=+1443.948960302" lastFinishedPulling="2025-12-03 13:33:21.433501456 +0000 UTC m=+1447.414421889" observedRunningTime="2025-12-03 13:33:26.208923627 +0000 UTC m=+1452.189844080" watchObservedRunningTime="2025-12-03 13:33:26.214255584 +0000 UTC m=+1452.195176017" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.250137 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-httpd-run\") pod \"105e22f6-36c9-4ab3-8e64-9f056a266704\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.250356 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"105e22f6-36c9-4ab3-8e64-9f056a266704\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.250811 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "105e22f6-36c9-4ab3-8e64-9f056a266704" (UID: "105e22f6-36c9-4ab3-8e64-9f056a266704"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.251520 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-internal-tls-certs\") pod \"105e22f6-36c9-4ab3-8e64-9f056a266704\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.251556 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-scripts\") pod \"105e22f6-36c9-4ab3-8e64-9f056a266704\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.251589 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-combined-ca-bundle\") pod \"105e22f6-36c9-4ab3-8e64-9f056a266704\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.251676 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-logs\") pod \"105e22f6-36c9-4ab3-8e64-9f056a266704\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.251706 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf7bs\" (UniqueName: \"kubernetes.io/projected/105e22f6-36c9-4ab3-8e64-9f056a266704-kube-api-access-vf7bs\") pod \"105e22f6-36c9-4ab3-8e64-9f056a266704\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.251743 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-config-data\") pod \"105e22f6-36c9-4ab3-8e64-9f056a266704\" (UID: \"105e22f6-36c9-4ab3-8e64-9f056a266704\") " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.252708 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-logs" (OuterVolumeSpecName: "logs") pod "105e22f6-36c9-4ab3-8e64-9f056a266704" (UID: "105e22f6-36c9-4ab3-8e64-9f056a266704"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.253840 4690 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.255268 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/105e22f6-36c9-4ab3-8e64-9f056a266704-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.259707 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "105e22f6-36c9-4ab3-8e64-9f056a266704" (UID: "105e22f6-36c9-4ab3-8e64-9f056a266704"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.272910 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-scripts" (OuterVolumeSpecName: "scripts") pod "105e22f6-36c9-4ab3-8e64-9f056a266704" (UID: "105e22f6-36c9-4ab3-8e64-9f056a266704"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.274460 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/105e22f6-36c9-4ab3-8e64-9f056a266704-kube-api-access-vf7bs" (OuterVolumeSpecName: "kube-api-access-vf7bs") pod "105e22f6-36c9-4ab3-8e64-9f056a266704" (UID: "105e22f6-36c9-4ab3-8e64-9f056a266704"). InnerVolumeSpecName "kube-api-access-vf7bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.357039 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "105e22f6-36c9-4ab3-8e64-9f056a266704" (UID: "105e22f6-36c9-4ab3-8e64-9f056a266704"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.360884 4690 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.360931 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.360964 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.360987 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf7bs\" (UniqueName: \"kubernetes.io/projected/105e22f6-36c9-4ab3-8e64-9f056a266704-kube-api-access-vf7bs\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.416706 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "105e22f6-36c9-4ab3-8e64-9f056a266704" (UID: "105e22f6-36c9-4ab3-8e64-9f056a266704"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.430645 4690 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.444065 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-config-data" (OuterVolumeSpecName: "config-data") pod "105e22f6-36c9-4ab3-8e64-9f056a266704" (UID: "105e22f6-36c9-4ab3-8e64-9f056a266704"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.462963 4690 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.463009 4690 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:26 crc kubenswrapper[4690]: I1203 13:33:26.463023 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/105e22f6-36c9-4ab3-8e64-9f056a266704-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.186675 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.186677 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"105e22f6-36c9-4ab3-8e64-9f056a266704","Type":"ContainerDied","Data":"a137ccf12ffa8a24112e602c0772b929c1098bb353882e08f88a421e0461a011"} Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.187216 4690 scope.go:117] "RemoveContainer" containerID="50fb8514b26b86d905c80d505bfd3f7977b619a006d3c4bd64746853de8bcb53" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.188572 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33ca8c1-d71d-4624-a362-f845fda14976","Type":"ContainerStarted","Data":"b4439c20e9feb7f5c8f5c10eaf29129f067af3c8fe23c34d8c392262a9130da1"} Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.190529 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ada5ac2f-e8b5-4437-b639-183fe2bf3c09","Type":"ContainerStarted","Data":"270b38dd57d1a37f66c2e6cf5be97edb876e4ad3dd1609c5f1adb560a31a8d1c"} Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.213977 4690 scope.go:117] "RemoveContainer" containerID="49017db7e96374547b2aab24e42d9d98a9647f40c8f34a5ef74506765384840c" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.219233 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.219210193 podStartE2EDuration="5.219210193s" podCreationTimestamp="2025-12-03 13:33:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:27.214554483 +0000 UTC m=+1453.195474936" watchObservedRunningTime="2025-12-03 13:33:27.219210193 +0000 UTC m=+1453.200130626" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.242626 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.260552 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.270762 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.272480 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:33:27 crc kubenswrapper[4690]: E1203 13:33:27.273024 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerName="glance-log" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.273040 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerName="glance-log" Dec 03 13:33:27 crc kubenswrapper[4690]: E1203 13:33:27.273053 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerName="glance-httpd" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.273060 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerName="glance-httpd" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.273264 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerName="glance-log" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.273284 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" containerName="glance-httpd" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.274487 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.277072 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.277306 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.284384 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.383842 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.383939 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f2638db-cd46-439c-bf63-4cd912784601-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.383993 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f2638db-cd46-439c-bf63-4cd912784601-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.384030 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.384057 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwldm\" (UniqueName: \"kubernetes.io/projected/8f2638db-cd46-439c-bf63-4cd912784601-kube-api-access-gwldm\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.384089 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.384251 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.384359 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.432264 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.491103 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.491187 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.491225 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f2638db-cd46-439c-bf63-4cd912784601-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.491261 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f2638db-cd46-439c-bf63-4cd912784601-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.491288 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.491317 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwldm\" (UniqueName: \"kubernetes.io/projected/8f2638db-cd46-439c-bf63-4cd912784601-kube-api-access-gwldm\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.491349 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.491458 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.502864 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f2638db-cd46-439c-bf63-4cd912784601-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.503577 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.504710 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-l66qt"] Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.505315 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f2638db-cd46-439c-bf63-4cd912784601-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.505437 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" podUID="14453e43-831f-4ea4-9a7f-57621e3999a4" containerName="dnsmasq-dns" containerID="cri-o://96163939beeee95229d9a3343b458be1a7d7e08afe9b3b9a31a7899ea09a9d1e" gracePeriod=10 Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.507807 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.510492 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.511783 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.525103 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2638db-cd46-439c-bf63-4cd912784601-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.546321 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.547104 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwldm\" (UniqueName: \"kubernetes.io/projected/8f2638db-cd46-439c-bf63-4cd912784601-kube-api-access-gwldm\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.552500 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f2638db-cd46-439c-bf63-4cd912784601\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:33:27 crc kubenswrapper[4690]: I1203 13:33:27.601789 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.207295 4690 generic.go:334] "Generic (PLEG): container finished" podID="14453e43-831f-4ea4-9a7f-57621e3999a4" containerID="96163939beeee95229d9a3343b458be1a7d7e08afe9b3b9a31a7899ea09a9d1e" exitCode=0 Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.207371 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" event={"ID":"14453e43-831f-4ea4-9a7f-57621e3999a4","Type":"ContainerDied","Data":"96163939beeee95229d9a3343b458be1a7d7e08afe9b3b9a31a7899ea09a9d1e"} Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.216385 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d33ca8c1-d71d-4624-a362-f845fda14976","Type":"ContainerStarted","Data":"ba8787ae1b33f110d744fa19017ddf31cfee19ef6f56da5fb8824bbfd72a012c"} Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.219075 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.305193 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.305736 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.305709822 podStartE2EDuration="4.305709822s" podCreationTimestamp="2025-12-03 13:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:28.287780188 +0000 UTC m=+1454.268700631" watchObservedRunningTime="2025-12-03 13:33:28.305709822 +0000 UTC m=+1454.286630255" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.325808 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="105e22f6-36c9-4ab3-8e64-9f056a266704" path="/var/lib/kubelet/pods/105e22f6-36c9-4ab3-8e64-9f056a266704/volumes" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.339478 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:33:28 crc kubenswrapper[4690]: W1203 13:33:28.342941 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f2638db_cd46_439c_bf63_4cd912784601.slice/crio-749c28657962acbd25f008f57299b453e29479d414d03aec78bace315bdb99db WatchSource:0}: Error finding container 749c28657962acbd25f008f57299b453e29479d414d03aec78bace315bdb99db: Status 404 returned error can't find the container with id 749c28657962acbd25f008f57299b453e29479d414d03aec78bace315bdb99db Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.704320 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.819930 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-nb\") pod \"14453e43-831f-4ea4-9a7f-57621e3999a4\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.820038 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-swift-storage-0\") pod \"14453e43-831f-4ea4-9a7f-57621e3999a4\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.820089 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-config\") pod \"14453e43-831f-4ea4-9a7f-57621e3999a4\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.820108 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-svc\") pod \"14453e43-831f-4ea4-9a7f-57621e3999a4\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.820207 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnh6j\" (UniqueName: \"kubernetes.io/projected/14453e43-831f-4ea4-9a7f-57621e3999a4-kube-api-access-rnh6j\") pod \"14453e43-831f-4ea4-9a7f-57621e3999a4\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.820316 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-sb\") pod \"14453e43-831f-4ea4-9a7f-57621e3999a4\" (UID: \"14453e43-831f-4ea4-9a7f-57621e3999a4\") " Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.829413 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14453e43-831f-4ea4-9a7f-57621e3999a4-kube-api-access-rnh6j" (OuterVolumeSpecName: "kube-api-access-rnh6j") pod "14453e43-831f-4ea4-9a7f-57621e3999a4" (UID: "14453e43-831f-4ea4-9a7f-57621e3999a4"). InnerVolumeSpecName "kube-api-access-rnh6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.882714 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14453e43-831f-4ea4-9a7f-57621e3999a4" (UID: "14453e43-831f-4ea4-9a7f-57621e3999a4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.891261 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14453e43-831f-4ea4-9a7f-57621e3999a4" (UID: "14453e43-831f-4ea4-9a7f-57621e3999a4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.901625 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "14453e43-831f-4ea4-9a7f-57621e3999a4" (UID: "14453e43-831f-4ea4-9a7f-57621e3999a4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.902955 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-config" (OuterVolumeSpecName: "config") pod "14453e43-831f-4ea4-9a7f-57621e3999a4" (UID: "14453e43-831f-4ea4-9a7f-57621e3999a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.905310 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14453e43-831f-4ea4-9a7f-57621e3999a4" (UID: "14453e43-831f-4ea4-9a7f-57621e3999a4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.923740 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.923789 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.923804 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.923818 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.923834 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14453e43-831f-4ea4-9a7f-57621e3999a4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:28 crc kubenswrapper[4690]: I1203 13:33:28.923844 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnh6j\" (UniqueName: \"kubernetes.io/projected/14453e43-831f-4ea4-9a7f-57621e3999a4-kube-api-access-rnh6j\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.228318 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" event={"ID":"14453e43-831f-4ea4-9a7f-57621e3999a4","Type":"ContainerDied","Data":"915f78d71448b06513375144c0f2219d6cf16634b09897024f6c79ad11c7c4df"} Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.228395 4690 scope.go:117] "RemoveContainer" containerID="96163939beeee95229d9a3343b458be1a7d7e08afe9b3b9a31a7899ea09a9d1e" Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.228389 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-l66qt" Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.232152 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2638db-cd46-439c-bf63-4cd912784601","Type":"ContainerStarted","Data":"749c28657962acbd25f008f57299b453e29479d414d03aec78bace315bdb99db"} Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.232448 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" containerName="cinder-scheduler" containerID="cri-o://4ba06fc7351cb8d2092580e2d1d94aa907ad00d987372a1856d65b3476db9555" gracePeriod=30 Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.232744 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" containerName="probe" containerID="cri-o://0f8d67945b6f649416d68ba70fffbb38afe567ab4b4c2d82d166e064f69a7b04" gracePeriod=30 Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.266559 4690 scope.go:117] "RemoveContainer" containerID="dbd56c23b40d664452b75f43ff085eb3ab84597759cf3f2d088d91805c35ef7d" Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.273899 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-l66qt"] Dec 03 13:33:29 crc kubenswrapper[4690]: I1203 13:33:29.283911 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-l66qt"] Dec 03 13:33:30 crc kubenswrapper[4690]: I1203 13:33:30.256053 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2638db-cd46-439c-bf63-4cd912784601","Type":"ContainerStarted","Data":"4dbff0e487f8a48ec42ed299a2e994a92c8ec715baad07cf8ddc104acd339a94"} Dec 03 13:33:30 crc kubenswrapper[4690]: I1203 13:33:30.341359 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14453e43-831f-4ea4-9a7f-57621e3999a4" path="/var/lib/kubelet/pods/14453e43-831f-4ea4-9a7f-57621e3999a4/volumes" Dec 03 13:33:30 crc kubenswrapper[4690]: E1203 13:33:30.834237 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode23c2dcc_ec98_49a4_92e9_db075d8f6f42.slice/crio-conmon-31ac604b3b9973bd8e383dc69aaff6758e2692c5d90e7d09a2b2aa1a2b830605.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.267803 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerID="31ac604b3b9973bd8e383dc69aaff6758e2692c5d90e7d09a2b2aa1a2b830605" exitCode=0 Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.267878 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerDied","Data":"31ac604b3b9973bd8e383dc69aaff6758e2692c5d90e7d09a2b2aa1a2b830605"} Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.270231 4690 generic.go:334] "Generic (PLEG): container finished" podID="fe83b063-4c0b-4a16-8460-da53de551d17" containerID="0f8d67945b6f649416d68ba70fffbb38afe567ab4b4c2d82d166e064f69a7b04" exitCode=0 Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.270256 4690 generic.go:334] "Generic (PLEG): container finished" podID="fe83b063-4c0b-4a16-8460-da53de551d17" containerID="4ba06fc7351cb8d2092580e2d1d94aa907ad00d987372a1856d65b3476db9555" exitCode=0 Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.270304 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe83b063-4c0b-4a16-8460-da53de551d17","Type":"ContainerDied","Data":"0f8d67945b6f649416d68ba70fffbb38afe567ab4b4c2d82d166e064f69a7b04"} Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.270331 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe83b063-4c0b-4a16-8460-da53de551d17","Type":"ContainerDied","Data":"4ba06fc7351cb8d2092580e2d1d94aa907ad00d987372a1856d65b3476db9555"} Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.272620 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2638db-cd46-439c-bf63-4cd912784601","Type":"ContainerStarted","Data":"8cd3ba1c195e1a5a19b935439c64006a37f61d1c05485aa35ea684098bcc5bdc"} Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.320783 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.320744192 podStartE2EDuration="4.320744192s" podCreationTimestamp="2025-12-03 13:33:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:31.305263521 +0000 UTC m=+1457.286183954" watchObservedRunningTime="2025-12-03 13:33:31.320744192 +0000 UTC m=+1457.301664625" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.669421 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.793491 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data\") pod \"fe83b063-4c0b-4a16-8460-da53de551d17\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.793533 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe83b063-4c0b-4a16-8460-da53de551d17-etc-machine-id\") pod \"fe83b063-4c0b-4a16-8460-da53de551d17\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.793587 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-combined-ca-bundle\") pod \"fe83b063-4c0b-4a16-8460-da53de551d17\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.793668 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-scripts\") pod \"fe83b063-4c0b-4a16-8460-da53de551d17\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.793712 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k55tb\" (UniqueName: \"kubernetes.io/projected/fe83b063-4c0b-4a16-8460-da53de551d17-kube-api-access-k55tb\") pod \"fe83b063-4c0b-4a16-8460-da53de551d17\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.793726 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fe83b063-4c0b-4a16-8460-da53de551d17-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fe83b063-4c0b-4a16-8460-da53de551d17" (UID: "fe83b063-4c0b-4a16-8460-da53de551d17"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.793803 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data-custom\") pod \"fe83b063-4c0b-4a16-8460-da53de551d17\" (UID: \"fe83b063-4c0b-4a16-8460-da53de551d17\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.794370 4690 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe83b063-4c0b-4a16-8460-da53de551d17-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.799412 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fe83b063-4c0b-4a16-8460-da53de551d17" (UID: "fe83b063-4c0b-4a16-8460-da53de551d17"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.799600 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe83b063-4c0b-4a16-8460-da53de551d17-kube-api-access-k55tb" (OuterVolumeSpecName: "kube-api-access-k55tb") pod "fe83b063-4c0b-4a16-8460-da53de551d17" (UID: "fe83b063-4c0b-4a16-8460-da53de551d17"). InnerVolumeSpecName "kube-api-access-k55tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.809799 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-scripts" (OuterVolumeSpecName: "scripts") pod "fe83b063-4c0b-4a16-8460-da53de551d17" (UID: "fe83b063-4c0b-4a16-8460-da53de551d17"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.866025 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe83b063-4c0b-4a16-8460-da53de551d17" (UID: "fe83b063-4c0b-4a16-8460-da53de551d17"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.886387 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.896308 4690 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.896347 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.896362 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.896372 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k55tb\" (UniqueName: \"kubernetes.io/projected/fe83b063-4c0b-4a16-8460-da53de551d17-kube-api-access-k55tb\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.912619 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data" (OuterVolumeSpecName: "config-data") pod "fe83b063-4c0b-4a16-8460-da53de551d17" (UID: "fe83b063-4c0b-4a16-8460-da53de551d17"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.997363 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-scripts\") pod \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.997411 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdzj7\" (UniqueName: \"kubernetes.io/projected/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-kube-api-access-wdzj7\") pod \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.997549 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-config-data\") pod \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.997571 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-sg-core-conf-yaml\") pod \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.997596 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-run-httpd\") pod \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.997635 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-log-httpd\") pod \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.997696 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-combined-ca-bundle\") pod \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\" (UID: \"e23c2dcc-ec98-49a4-92e9-db075d8f6f42\") " Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.998169 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe83b063-4c0b-4a16-8460-da53de551d17-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.999105 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e23c2dcc-ec98-49a4-92e9-db075d8f6f42" (UID: "e23c2dcc-ec98-49a4-92e9-db075d8f6f42"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:31 crc kubenswrapper[4690]: I1203 13:33:31.999449 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e23c2dcc-ec98-49a4-92e9-db075d8f6f42" (UID: "e23c2dcc-ec98-49a4-92e9-db075d8f6f42"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.002781 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-scripts" (OuterVolumeSpecName: "scripts") pod "e23c2dcc-ec98-49a4-92e9-db075d8f6f42" (UID: "e23c2dcc-ec98-49a4-92e9-db075d8f6f42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.003984 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-kube-api-access-wdzj7" (OuterVolumeSpecName: "kube-api-access-wdzj7") pod "e23c2dcc-ec98-49a4-92e9-db075d8f6f42" (UID: "e23c2dcc-ec98-49a4-92e9-db075d8f6f42"). InnerVolumeSpecName "kube-api-access-wdzj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.038073 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e23c2dcc-ec98-49a4-92e9-db075d8f6f42" (UID: "e23c2dcc-ec98-49a4-92e9-db075d8f6f42"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.082027 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e23c2dcc-ec98-49a4-92e9-db075d8f6f42" (UID: "e23c2dcc-ec98-49a4-92e9-db075d8f6f42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.100187 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.100226 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.100239 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdzj7\" (UniqueName: \"kubernetes.io/projected/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-kube-api-access-wdzj7\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.100251 4690 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.100259 4690 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.100267 4690 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.115546 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-config-data" (OuterVolumeSpecName: "config-data") pod "e23c2dcc-ec98-49a4-92e9-db075d8f6f42" (UID: "e23c2dcc-ec98-49a4-92e9-db075d8f6f42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.203697 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23c2dcc-ec98-49a4-92e9-db075d8f6f42-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.286794 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.287291 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e23c2dcc-ec98-49a4-92e9-db075d8f6f42","Type":"ContainerDied","Data":"8c342cd85095f4d239f03185c9c6245d140cb5aa22c6d1050fcdfd9748ae4fb5"} Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.287617 4690 scope.go:117] "RemoveContainer" containerID="9c6f89be5976ed0d34f8ab2aa5e740da117e870e128be0d0bab58fa754224aff" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.292447 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe83b063-4c0b-4a16-8460-da53de551d17","Type":"ContainerDied","Data":"83e8f690e6cbd43bfd04b810eb0beba9d22b0d73408e08032425ef03dd17b51e"} Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.292627 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.323552 4690 scope.go:117] "RemoveContainer" containerID="6ae82468b738e6a64295012886490c63cc488a690beeb949309d80e27b75f266" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.345696 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.371996 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.376555 4690 scope.go:117] "RemoveContainer" containerID="31ac604b3b9973bd8e383dc69aaff6758e2692c5d90e7d09a2b2aa1a2b830605" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.384570 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.396931 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:32 crc kubenswrapper[4690]: E1203 13:33:32.397633 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="ceilometer-notification-agent" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.397705 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="ceilometer-notification-agent" Dec 03 13:33:32 crc kubenswrapper[4690]: E1203 13:33:32.397770 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14453e43-831f-4ea4-9a7f-57621e3999a4" containerName="init" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.397837 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="14453e43-831f-4ea4-9a7f-57621e3999a4" containerName="init" Dec 03 13:33:32 crc kubenswrapper[4690]: E1203 13:33:32.397938 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14453e43-831f-4ea4-9a7f-57621e3999a4" containerName="dnsmasq-dns" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.397998 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="14453e43-831f-4ea4-9a7f-57621e3999a4" containerName="dnsmasq-dns" Dec 03 13:33:32 crc kubenswrapper[4690]: E1203 13:33:32.398057 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="proxy-httpd" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.398110 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="proxy-httpd" Dec 03 13:33:32 crc kubenswrapper[4690]: E1203 13:33:32.398166 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" containerName="probe" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.398220 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" containerName="probe" Dec 03 13:33:32 crc kubenswrapper[4690]: E1203 13:33:32.398283 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="sg-core" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.400809 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="sg-core" Dec 03 13:33:32 crc kubenswrapper[4690]: E1203 13:33:32.400969 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="ceilometer-central-agent" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.401033 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="ceilometer-central-agent" Dec 03 13:33:32 crc kubenswrapper[4690]: E1203 13:33:32.401087 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" containerName="cinder-scheduler" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.401142 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" containerName="cinder-scheduler" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.401634 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" containerName="probe" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.401715 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="ceilometer-central-agent" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.401781 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="14453e43-831f-4ea4-9a7f-57621e3999a4" containerName="dnsmasq-dns" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.401830 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="sg-core" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.401915 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="ceilometer-notification-agent" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.401977 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" containerName="cinder-scheduler" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.402042 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" containerName="proxy-httpd" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.408045 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.410394 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.414454 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.414609 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.425083 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.429853 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.432994 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.439847 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.454531 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.460077 4690 scope.go:117] "RemoveContainer" containerID="294b0b57fa1be9e729837ce3dda9c0dff30aab373239cb0aacc2f8d55c41747c" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.488714 4690 scope.go:117] "RemoveContainer" containerID="0f8d67945b6f649416d68ba70fffbb38afe567ab4b4c2d82d166e064f69a7b04" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.509763 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.509813 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.509833 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.509879 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-config-data\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.509942 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdd4c\" (UniqueName: \"kubernetes.io/projected/87b18c03-5cae-4ee6-9634-f3d832d57d66-kube-api-access-mdd4c\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.509961 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-scripts\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.510037 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-log-httpd\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.510086 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-scripts\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.510131 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm2r8\" (UniqueName: \"kubernetes.io/projected/fa834771-9deb-40da-b5ae-4f8d60667d2b-kube-api-access-qm2r8\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.510151 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-config-data\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.510172 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fa834771-9deb-40da-b5ae-4f8d60667d2b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.510195 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.510220 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-run-httpd\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.513392 4690 scope.go:117] "RemoveContainer" containerID="4ba06fc7351cb8d2092580e2d1d94aa907ad00d987372a1856d65b3476db9555" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612075 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-log-httpd\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612178 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-scripts\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612246 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm2r8\" (UniqueName: \"kubernetes.io/projected/fa834771-9deb-40da-b5ae-4f8d60667d2b-kube-api-access-qm2r8\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612273 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-config-data\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612304 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fa834771-9deb-40da-b5ae-4f8d60667d2b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612332 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612355 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-run-httpd\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612403 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fa834771-9deb-40da-b5ae-4f8d60667d2b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612410 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612500 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612528 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612565 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-config-data\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612657 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdd4c\" (UniqueName: \"kubernetes.io/projected/87b18c03-5cae-4ee6-9634-f3d832d57d66-kube-api-access-mdd4c\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.612689 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-scripts\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.613364 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-run-httpd\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.613651 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-log-httpd\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.618309 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.618712 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-scripts\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.618770 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.619380 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-config-data\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.619829 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.628684 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-scripts\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.630193 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.630519 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa834771-9deb-40da-b5ae-4f8d60667d2b-config-data\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.632720 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdd4c\" (UniqueName: \"kubernetes.io/projected/87b18c03-5cae-4ee6-9634-f3d832d57d66-kube-api-access-mdd4c\") pod \"ceilometer-0\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.633355 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm2r8\" (UniqueName: \"kubernetes.io/projected/fa834771-9deb-40da-b5ae-4f8d60667d2b-kube-api-access-qm2r8\") pod \"cinder-scheduler-0\" (UID: \"fa834771-9deb-40da-b5ae-4f8d60667d2b\") " pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.727279 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.727340 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.742501 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.766075 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.767519 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:33:32 crc kubenswrapper[4690]: I1203 13:33:32.770206 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.275721 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.310892 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerStarted","Data":"bb858caaf6787b0052a61b977cc1e28fc7464859fc635052df99513767ca2431"} Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.310973 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.310996 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.349682 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:33:33 crc kubenswrapper[4690]: W1203 13:33:33.359879 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa834771_9deb_40da_b5ae_4f8d60667d2b.slice/crio-d0274af914aa824400ed83d08dcd9e6c0acd25db1e37eecd1dc471ddcbe2c891 WatchSource:0}: Error finding container d0274af914aa824400ed83d08dcd9e6c0acd25db1e37eecd1dc471ddcbe2c891: Status 404 returned error can't find the container with id d0274af914aa824400ed83d08dcd9e6c0acd25db1e37eecd1dc471ddcbe2c891 Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.394331 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pzvps"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.396340 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.435394 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pzvps"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.539312 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc4hc\" (UniqueName: \"kubernetes.io/projected/2478ef86-8c11-4d36-88b1-eea0a3b2310a-kube-api-access-jc4hc\") pod \"nova-api-db-create-pzvps\" (UID: \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\") " pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.539408 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2478ef86-8c11-4d36-88b1-eea0a3b2310a-operator-scripts\") pod \"nova-api-db-create-pzvps\" (UID: \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\") " pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.546092 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-8k487"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.549042 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.555086 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-8k487"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.631957 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8cf5-account-create-update-2m8zh"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.634048 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.645262 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.646749 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gxfw\" (UniqueName: \"kubernetes.io/projected/dbddf981-d575-4517-b0e6-f494b9a92fe6-kube-api-access-6gxfw\") pod \"nova-cell0-db-create-8k487\" (UID: \"dbddf981-d575-4517-b0e6-f494b9a92fe6\") " pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.646857 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2478ef86-8c11-4d36-88b1-eea0a3b2310a-operator-scripts\") pod \"nova-api-db-create-pzvps\" (UID: \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\") " pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.646954 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbddf981-d575-4517-b0e6-f494b9a92fe6-operator-scripts\") pod \"nova-cell0-db-create-8k487\" (UID: \"dbddf981-d575-4517-b0e6-f494b9a92fe6\") " pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.651165 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc4hc\" (UniqueName: \"kubernetes.io/projected/2478ef86-8c11-4d36-88b1-eea0a3b2310a-kube-api-access-jc4hc\") pod \"nova-api-db-create-pzvps\" (UID: \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\") " pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.652672 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2478ef86-8c11-4d36-88b1-eea0a3b2310a-operator-scripts\") pod \"nova-api-db-create-pzvps\" (UID: \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\") " pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.671086 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8cf5-account-create-update-2m8zh"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.709722 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-rxz2w"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.711706 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.730305 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc4hc\" (UniqueName: \"kubernetes.io/projected/2478ef86-8c11-4d36-88b1-eea0a3b2310a-kube-api-access-jc4hc\") pod \"nova-api-db-create-pzvps\" (UID: \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\") " pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.737483 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.753459 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbddf981-d575-4517-b0e6-f494b9a92fe6-operator-scripts\") pod \"nova-cell0-db-create-8k487\" (UID: \"dbddf981-d575-4517-b0e6-f494b9a92fe6\") " pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.753575 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-operator-scripts\") pod \"nova-api-8cf5-account-create-update-2m8zh\" (UID: \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\") " pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.753629 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qm8t\" (UniqueName: \"kubernetes.io/projected/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-kube-api-access-9qm8t\") pod \"nova-api-8cf5-account-create-update-2m8zh\" (UID: \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\") " pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.753666 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gxfw\" (UniqueName: \"kubernetes.io/projected/dbddf981-d575-4517-b0e6-f494b9a92fe6-kube-api-access-6gxfw\") pod \"nova-cell0-db-create-8k487\" (UID: \"dbddf981-d575-4517-b0e6-f494b9a92fe6\") " pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.754721 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbddf981-d575-4517-b0e6-f494b9a92fe6-operator-scripts\") pod \"nova-cell0-db-create-8k487\" (UID: \"dbddf981-d575-4517-b0e6-f494b9a92fe6\") " pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.754784 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rxz2w"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.779481 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gxfw\" (UniqueName: \"kubernetes.io/projected/dbddf981-d575-4517-b0e6-f494b9a92fe6-kube-api-access-6gxfw\") pod \"nova-cell0-db-create-8k487\" (UID: \"dbddf981-d575-4517-b0e6-f494b9a92fe6\") " pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.813649 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-de32-account-create-update-d5v2x"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.815327 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.819181 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.829085 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-de32-account-create-update-d5v2x"] Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.855554 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-operator-scripts\") pod \"nova-api-8cf5-account-create-update-2m8zh\" (UID: \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\") " pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.855626 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbwp2\" (UniqueName: \"kubernetes.io/projected/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-kube-api-access-hbwp2\") pod \"nova-cell1-db-create-rxz2w\" (UID: \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\") " pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.855657 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-operator-scripts\") pod \"nova-cell1-db-create-rxz2w\" (UID: \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\") " pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.855693 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qm8t\" (UniqueName: \"kubernetes.io/projected/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-kube-api-access-9qm8t\") pod \"nova-api-8cf5-account-create-update-2m8zh\" (UID: \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\") " pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.864671 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-operator-scripts\") pod \"nova-api-8cf5-account-create-update-2m8zh\" (UID: \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\") " pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.883464 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qm8t\" (UniqueName: \"kubernetes.io/projected/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-kube-api-access-9qm8t\") pod \"nova-api-8cf5-account-create-update-2m8zh\" (UID: \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\") " pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.963270 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-operator-scripts\") pod \"nova-cell0-de32-account-create-update-d5v2x\" (UID: \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\") " pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.968458 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbwp2\" (UniqueName: \"kubernetes.io/projected/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-kube-api-access-hbwp2\") pod \"nova-cell1-db-create-rxz2w\" (UID: \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\") " pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.968537 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-operator-scripts\") pod \"nova-cell1-db-create-rxz2w\" (UID: \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\") " pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.968742 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq5hl\" (UniqueName: \"kubernetes.io/projected/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-kube-api-access-kq5hl\") pod \"nova-cell0-de32-account-create-update-d5v2x\" (UID: \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\") " pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.971701 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-operator-scripts\") pod \"nova-cell1-db-create-rxz2w\" (UID: \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\") " pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.994555 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:33 crc kubenswrapper[4690]: I1203 13:33:33.994720 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.000403 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbwp2\" (UniqueName: \"kubernetes.io/projected/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-kube-api-access-hbwp2\") pod \"nova-cell1-db-create-rxz2w\" (UID: \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\") " pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.045027 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5b13-account-create-update-w7g9f"] Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.046796 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.049366 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.073089 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-operator-scripts\") pod \"nova-cell0-de32-account-create-update-d5v2x\" (UID: \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\") " pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.073679 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq5hl\" (UniqueName: \"kubernetes.io/projected/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-kube-api-access-kq5hl\") pod \"nova-cell0-de32-account-create-update-d5v2x\" (UID: \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\") " pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.074127 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-operator-scripts\") pod \"nova-cell0-de32-account-create-update-d5v2x\" (UID: \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\") " pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.074813 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5b13-account-create-update-w7g9f"] Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.097334 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq5hl\" (UniqueName: \"kubernetes.io/projected/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-kube-api-access-kq5hl\") pod \"nova-cell0-de32-account-create-update-d5v2x\" (UID: \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\") " pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.101024 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.176086 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8137b9f0-e7c6-4134-ae61-5599982fa141-operator-scripts\") pod \"nova-cell1-5b13-account-create-update-w7g9f\" (UID: \"8137b9f0-e7c6-4134-ae61-5599982fa141\") " pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.176262 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7n6c\" (UniqueName: \"kubernetes.io/projected/8137b9f0-e7c6-4134-ae61-5599982fa141-kube-api-access-z7n6c\") pod \"nova-cell1-5b13-account-create-update-w7g9f\" (UID: \"8137b9f0-e7c6-4134-ae61-5599982fa141\") " pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.278862 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8137b9f0-e7c6-4134-ae61-5599982fa141-operator-scripts\") pod \"nova-cell1-5b13-account-create-update-w7g9f\" (UID: \"8137b9f0-e7c6-4134-ae61-5599982fa141\") " pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.279091 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7n6c\" (UniqueName: \"kubernetes.io/projected/8137b9f0-e7c6-4134-ae61-5599982fa141-kube-api-access-z7n6c\") pod \"nova-cell1-5b13-account-create-update-w7g9f\" (UID: \"8137b9f0-e7c6-4134-ae61-5599982fa141\") " pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.279855 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8137b9f0-e7c6-4134-ae61-5599982fa141-operator-scripts\") pod \"nova-cell1-5b13-account-create-update-w7g9f\" (UID: \"8137b9f0-e7c6-4134-ae61-5599982fa141\") " pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.296198 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.302648 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7n6c\" (UniqueName: \"kubernetes.io/projected/8137b9f0-e7c6-4134-ae61-5599982fa141-kube-api-access-z7n6c\") pod \"nova-cell1-5b13-account-create-update-w7g9f\" (UID: \"8137b9f0-e7c6-4134-ae61-5599982fa141\") " pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.367644 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e23c2dcc-ec98-49a4-92e9-db075d8f6f42" path="/var/lib/kubelet/pods/e23c2dcc-ec98-49a4-92e9-db075d8f6f42/volumes" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.368439 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe83b063-4c0b-4a16-8460-da53de551d17" path="/var/lib/kubelet/pods/fe83b063-4c0b-4a16-8460-da53de551d17/volumes" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.369275 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fa834771-9deb-40da-b5ae-4f8d60667d2b","Type":"ContainerStarted","Data":"d0274af914aa824400ed83d08dcd9e6c0acd25db1e37eecd1dc471ddcbe2c891"} Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.381263 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerStarted","Data":"3e52b546c1b666c3e15e4a412825dee0bc44d343e3dbc9732b55d0542bac5216"} Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.461461 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pzvps"] Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.471434 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.776133 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8cf5-account-create-update-2m8zh"] Dec 03 13:33:34 crc kubenswrapper[4690]: I1203 13:33:34.980422 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-8k487"] Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.186346 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-de32-account-create-update-d5v2x"] Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.220244 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rxz2w"] Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.437226 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pzvps" event={"ID":"2478ef86-8c11-4d36-88b1-eea0a3b2310a","Type":"ContainerStarted","Data":"6a91b52845253aef44e053f1c33a5981cee01a273677f375c5fc8ff8e7539d3d"} Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.437295 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pzvps" event={"ID":"2478ef86-8c11-4d36-88b1-eea0a3b2310a","Type":"ContainerStarted","Data":"6deaa4f0dddb21bae0930b021102d63c11e77ffe265173987b129fdb380c690a"} Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.452140 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8k487" event={"ID":"dbddf981-d575-4517-b0e6-f494b9a92fe6","Type":"ContainerStarted","Data":"7dcceda78e8db08de4a1dd7dfa098faee87f1a6f1fe0720f7ed916fded756347"} Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.452580 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5b13-account-create-update-w7g9f"] Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.467137 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" event={"ID":"5bf7fd9d-669f-4d15-a432-8b796d5f1de5","Type":"ContainerStarted","Data":"6bf3767fcb7fecb21d8044b1098141d095e4ece9a92034331a259c18b01804fd"} Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.486564 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rxz2w" event={"ID":"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3","Type":"ContainerStarted","Data":"26684a36ca392912a8b6d2276ea3aac82d50dc7bb0e5d48797845ccfccc38f4d"} Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.505657 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" event={"ID":"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8","Type":"ContainerStarted","Data":"523043849ff1f4c50e29decc406ed18ed8b4abb0964daa34bd079b455ba7932f"} Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.516808 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.516833 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:33:35 crc kubenswrapper[4690]: I1203 13:33:35.517767 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fa834771-9deb-40da-b5ae-4f8d60667d2b","Type":"ContainerStarted","Data":"02dd344024b9dec1da71ef1b085eca9958b034f360a6166ab9d4b1f3f32372fa"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.330119 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.528620 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fa834771-9deb-40da-b5ae-4f8d60667d2b","Type":"ContainerStarted","Data":"f0252e3e84e29070739d7fa0eaee1b63a0717e7b2818fee48e5227dff3849615"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.531984 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" event={"ID":"8137b9f0-e7c6-4134-ae61-5599982fa141","Type":"ContainerStarted","Data":"6e249d4455c2ded66f572068a6a54d1a5c7e0e1effa4e253c7fd6e290b7d49c2"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.532020 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" event={"ID":"8137b9f0-e7c6-4134-ae61-5599982fa141","Type":"ContainerStarted","Data":"58fc7e1d95bb2848c5269f93cab1485679970b99402588c6bf5991043295a0bf"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.535911 4690 generic.go:334] "Generic (PLEG): container finished" podID="2478ef86-8c11-4d36-88b1-eea0a3b2310a" containerID="6a91b52845253aef44e053f1c33a5981cee01a273677f375c5fc8ff8e7539d3d" exitCode=0 Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.535977 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pzvps" event={"ID":"2478ef86-8c11-4d36-88b1-eea0a3b2310a","Type":"ContainerDied","Data":"6a91b52845253aef44e053f1c33a5981cee01a273677f375c5fc8ff8e7539d3d"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.538772 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8k487" event={"ID":"dbddf981-d575-4517-b0e6-f494b9a92fe6","Type":"ContainerStarted","Data":"2a92c73f8d3c14ce13f0f46cd1e56c15be1d86522956396c59edcce9221675b4"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.543099 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerStarted","Data":"823c62f78dfd2ed299db0b1cd0ac3f4f75381387f193254c72c3ced9000479e1"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.549420 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" event={"ID":"5bf7fd9d-669f-4d15-a432-8b796d5f1de5","Type":"ContainerStarted","Data":"f82022919bfa95568653fc71076b34c6a96a848f2fd2ebd9e87bb884d122a26b"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.555373 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rxz2w" event={"ID":"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3","Type":"ContainerStarted","Data":"79843863f21b26dc627742ae974accc39c1d70ce7eb185efa682ff1e93cd2def"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.563148 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.563225 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" event={"ID":"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8","Type":"ContainerStarted","Data":"26f89748c95d5f3a2f242b46d57fe3a6c937a90a0140641b5f163e93f90407d7"} Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.568102 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.568079855 podStartE2EDuration="4.568079855s" podCreationTimestamp="2025-12-03 13:33:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:36.558372394 +0000 UTC m=+1462.539292827" watchObservedRunningTime="2025-12-03 13:33:36.568079855 +0000 UTC m=+1462.549000288" Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.577191 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-8k487" podStartSLOduration=3.577173901 podStartE2EDuration="3.577173901s" podCreationTimestamp="2025-12-03 13:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:36.57638238 +0000 UTC m=+1462.557302823" watchObservedRunningTime="2025-12-03 13:33:36.577173901 +0000 UTC m=+1462.558094334" Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.611700 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" podStartSLOduration=3.611674093 podStartE2EDuration="3.611674093s" podCreationTimestamp="2025-12-03 13:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:36.593928014 +0000 UTC m=+1462.574848467" watchObservedRunningTime="2025-12-03 13:33:36.611674093 +0000 UTC m=+1462.592594526" Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.650110 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-rxz2w" podStartSLOduration=3.650084948 podStartE2EDuration="3.650084948s" podCreationTimestamp="2025-12-03 13:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:36.619715251 +0000 UTC m=+1462.600635694" watchObservedRunningTime="2025-12-03 13:33:36.650084948 +0000 UTC m=+1462.631005381" Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.666656 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" podStartSLOduration=3.666632366 podStartE2EDuration="3.666632366s" podCreationTimestamp="2025-12-03 13:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:36.63705591 +0000 UTC m=+1462.617976343" watchObservedRunningTime="2025-12-03 13:33:36.666632366 +0000 UTC m=+1462.647552809" Dec 03 13:33:36 crc kubenswrapper[4690]: I1203 13:33:36.676367 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" podStartSLOduration=3.676343917 podStartE2EDuration="3.676343917s" podCreationTimestamp="2025-12-03 13:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:33:36.650300493 +0000 UTC m=+1462.631220926" watchObservedRunningTime="2025-12-03 13:33:36.676343917 +0000 UTC m=+1462.657264350" Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.105615 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.599890 4690 generic.go:334] "Generic (PLEG): container finished" podID="ad6e77ea-b90d-4def-8ab1-d761fc41ebe3" containerID="79843863f21b26dc627742ae974accc39c1d70ce7eb185efa682ff1e93cd2def" exitCode=0 Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.599999 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rxz2w" event={"ID":"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3","Type":"ContainerDied","Data":"79843863f21b26dc627742ae974accc39c1d70ce7eb185efa682ff1e93cd2def"} Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.602443 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.602498 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.603301 4690 generic.go:334] "Generic (PLEG): container finished" podID="dbddf981-d575-4517-b0e6-f494b9a92fe6" containerID="2a92c73f8d3c14ce13f0f46cd1e56c15be1d86522956396c59edcce9221675b4" exitCode=0 Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.603499 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8k487" event={"ID":"dbddf981-d575-4517-b0e6-f494b9a92fe6","Type":"ContainerDied","Data":"2a92c73f8d3c14ce13f0f46cd1e56c15be1d86522956396c59edcce9221675b4"} Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.674362 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.699728 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.746653 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 13:33:37 crc kubenswrapper[4690]: I1203 13:33:37.799986 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.183504 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.342815 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc4hc\" (UniqueName: \"kubernetes.io/projected/2478ef86-8c11-4d36-88b1-eea0a3b2310a-kube-api-access-jc4hc\") pod \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\" (UID: \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\") " Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.343066 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2478ef86-8c11-4d36-88b1-eea0a3b2310a-operator-scripts\") pod \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\" (UID: \"2478ef86-8c11-4d36-88b1-eea0a3b2310a\") " Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.344038 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2478ef86-8c11-4d36-88b1-eea0a3b2310a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2478ef86-8c11-4d36-88b1-eea0a3b2310a" (UID: "2478ef86-8c11-4d36-88b1-eea0a3b2310a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.353390 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2478ef86-8c11-4d36-88b1-eea0a3b2310a-kube-api-access-jc4hc" (OuterVolumeSpecName: "kube-api-access-jc4hc") pod "2478ef86-8c11-4d36-88b1-eea0a3b2310a" (UID: "2478ef86-8c11-4d36-88b1-eea0a3b2310a"). InnerVolumeSpecName "kube-api-access-jc4hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.448160 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc4hc\" (UniqueName: \"kubernetes.io/projected/2478ef86-8c11-4d36-88b1-eea0a3b2310a-kube-api-access-jc4hc\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.448208 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2478ef86-8c11-4d36-88b1-eea0a3b2310a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.613593 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pzvps" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.613588 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pzvps" event={"ID":"2478ef86-8c11-4d36-88b1-eea0a3b2310a","Type":"ContainerDied","Data":"6deaa4f0dddb21bae0930b021102d63c11e77ffe265173987b129fdb380c690a"} Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.613723 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6deaa4f0dddb21bae0930b021102d63c11e77ffe265173987b129fdb380c690a" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.618230 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerStarted","Data":"760ccb4bdd5df2a779c960b37cd6966a39dcabae942397164f21d5347e8b3f3c"} Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.619510 4690 generic.go:334] "Generic (PLEG): container finished" podID="8137b9f0-e7c6-4134-ae61-5599982fa141" containerID="6e249d4455c2ded66f572068a6a54d1a5c7e0e1effa4e253c7fd6e290b7d49c2" exitCode=0 Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.619624 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" event={"ID":"8137b9f0-e7c6-4134-ae61-5599982fa141","Type":"ContainerDied","Data":"6e249d4455c2ded66f572068a6a54d1a5c7e0e1effa4e253c7fd6e290b7d49c2"} Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.620184 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:38 crc kubenswrapper[4690]: I1203 13:33:38.620212 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.321146 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.327582 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.469582 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbddf981-d575-4517-b0e6-f494b9a92fe6-operator-scripts\") pod \"dbddf981-d575-4517-b0e6-f494b9a92fe6\" (UID: \"dbddf981-d575-4517-b0e6-f494b9a92fe6\") " Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.470835 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbwp2\" (UniqueName: \"kubernetes.io/projected/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-kube-api-access-hbwp2\") pod \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\" (UID: \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\") " Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.470994 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-operator-scripts\") pod \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\" (UID: \"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3\") " Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.471096 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gxfw\" (UniqueName: \"kubernetes.io/projected/dbddf981-d575-4517-b0e6-f494b9a92fe6-kube-api-access-6gxfw\") pod \"dbddf981-d575-4517-b0e6-f494b9a92fe6\" (UID: \"dbddf981-d575-4517-b0e6-f494b9a92fe6\") " Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.470709 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbddf981-d575-4517-b0e6-f494b9a92fe6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dbddf981-d575-4517-b0e6-f494b9a92fe6" (UID: "dbddf981-d575-4517-b0e6-f494b9a92fe6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.472878 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad6e77ea-b90d-4def-8ab1-d761fc41ebe3" (UID: "ad6e77ea-b90d-4def-8ab1-d761fc41ebe3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.479071 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-kube-api-access-hbwp2" (OuterVolumeSpecName: "kube-api-access-hbwp2") pod "ad6e77ea-b90d-4def-8ab1-d761fc41ebe3" (UID: "ad6e77ea-b90d-4def-8ab1-d761fc41ebe3"). InnerVolumeSpecName "kube-api-access-hbwp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.489072 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbddf981-d575-4517-b0e6-f494b9a92fe6-kube-api-access-6gxfw" (OuterVolumeSpecName: "kube-api-access-6gxfw") pod "dbddf981-d575-4517-b0e6-f494b9a92fe6" (UID: "dbddf981-d575-4517-b0e6-f494b9a92fe6"). InnerVolumeSpecName "kube-api-access-6gxfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.574274 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbwp2\" (UniqueName: \"kubernetes.io/projected/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-kube-api-access-hbwp2\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.574321 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.574334 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gxfw\" (UniqueName: \"kubernetes.io/projected/dbddf981-d575-4517-b0e6-f494b9a92fe6-kube-api-access-6gxfw\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.574346 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbddf981-d575-4517-b0e6-f494b9a92fe6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.631229 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8k487" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.631231 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8k487" event={"ID":"dbddf981-d575-4517-b0e6-f494b9a92fe6","Type":"ContainerDied","Data":"7dcceda78e8db08de4a1dd7dfa098faee87f1a6f1fe0720f7ed916fded756347"} Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.631291 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dcceda78e8db08de4a1dd7dfa098faee87f1a6f1fe0720f7ed916fded756347" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.632960 4690 generic.go:334] "Generic (PLEG): container finished" podID="5bf7fd9d-669f-4d15-a432-8b796d5f1de5" containerID="f82022919bfa95568653fc71076b34c6a96a848f2fd2ebd9e87bb884d122a26b" exitCode=0 Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.632999 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" event={"ID":"5bf7fd9d-669f-4d15-a432-8b796d5f1de5","Type":"ContainerDied","Data":"f82022919bfa95568653fc71076b34c6a96a848f2fd2ebd9e87bb884d122a26b"} Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.634524 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rxz2w" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.634533 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rxz2w" event={"ID":"ad6e77ea-b90d-4def-8ab1-d761fc41ebe3","Type":"ContainerDied","Data":"26684a36ca392912a8b6d2276ea3aac82d50dc7bb0e5d48797845ccfccc38f4d"} Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.634590 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26684a36ca392912a8b6d2276ea3aac82d50dc7bb0e5d48797845ccfccc38f4d" Dec 03 13:33:39 crc kubenswrapper[4690]: I1203 13:33:39.999402 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.091451 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7n6c\" (UniqueName: \"kubernetes.io/projected/8137b9f0-e7c6-4134-ae61-5599982fa141-kube-api-access-z7n6c\") pod \"8137b9f0-e7c6-4134-ae61-5599982fa141\" (UID: \"8137b9f0-e7c6-4134-ae61-5599982fa141\") " Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.091535 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8137b9f0-e7c6-4134-ae61-5599982fa141-operator-scripts\") pod \"8137b9f0-e7c6-4134-ae61-5599982fa141\" (UID: \"8137b9f0-e7c6-4134-ae61-5599982fa141\") " Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.092973 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8137b9f0-e7c6-4134-ae61-5599982fa141-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8137b9f0-e7c6-4134-ae61-5599982fa141" (UID: "8137b9f0-e7c6-4134-ae61-5599982fa141"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.100519 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8137b9f0-e7c6-4134-ae61-5599982fa141-kube-api-access-z7n6c" (OuterVolumeSpecName: "kube-api-access-z7n6c") pod "8137b9f0-e7c6-4134-ae61-5599982fa141" (UID: "8137b9f0-e7c6-4134-ae61-5599982fa141"). InnerVolumeSpecName "kube-api-access-z7n6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.194186 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7n6c\" (UniqueName: \"kubernetes.io/projected/8137b9f0-e7c6-4134-ae61-5599982fa141-kube-api-access-z7n6c\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.194231 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8137b9f0-e7c6-4134-ae61-5599982fa141-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.645406 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.646173 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b13-account-create-update-w7g9f" event={"ID":"8137b9f0-e7c6-4134-ae61-5599982fa141","Type":"ContainerDied","Data":"58fc7e1d95bb2848c5269f93cab1485679970b99402588c6bf5991043295a0bf"} Dec 03 13:33:40 crc kubenswrapper[4690]: I1203 13:33:40.646196 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58fc7e1d95bb2848c5269f93cab1485679970b99402588c6bf5991043295a0bf" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.079398 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.215368 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qm8t\" (UniqueName: \"kubernetes.io/projected/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-kube-api-access-9qm8t\") pod \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\" (UID: \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\") " Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.215462 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-operator-scripts\") pod \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\" (UID: \"5bf7fd9d-669f-4d15-a432-8b796d5f1de5\") " Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.216357 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5bf7fd9d-669f-4d15-a432-8b796d5f1de5" (UID: "5bf7fd9d-669f-4d15-a432-8b796d5f1de5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.224726 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-kube-api-access-9qm8t" (OuterVolumeSpecName: "kube-api-access-9qm8t") pod "5bf7fd9d-669f-4d15-a432-8b796d5f1de5" (UID: "5bf7fd9d-669f-4d15-a432-8b796d5f1de5"). InnerVolumeSpecName "kube-api-access-9qm8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.318097 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.318130 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qm8t\" (UniqueName: \"kubernetes.io/projected/5bf7fd9d-669f-4d15-a432-8b796d5f1de5-kube-api-access-9qm8t\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.605428 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.605561 4690 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.606754 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.657986 4690 generic.go:334] "Generic (PLEG): container finished" podID="16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8" containerID="26f89748c95d5f3a2f242b46d57fe3a6c937a90a0140641b5f163e93f90407d7" exitCode=0 Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.658092 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" event={"ID":"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8","Type":"ContainerDied","Data":"26f89748c95d5f3a2f242b46d57fe3a6c937a90a0140641b5f163e93f90407d7"} Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.661034 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" event={"ID":"5bf7fd9d-669f-4d15-a432-8b796d5f1de5","Type":"ContainerDied","Data":"6bf3767fcb7fecb21d8044b1098141d095e4ece9a92034331a259c18b01804fd"} Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.661097 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8cf5-account-create-update-2m8zh" Dec 03 13:33:41 crc kubenswrapper[4690]: I1203 13:33:41.661098 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bf3767fcb7fecb21d8044b1098141d095e4ece9a92034331a259c18b01804fd" Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.097155 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.155539 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-operator-scripts\") pod \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\" (UID: \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\") " Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.155696 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq5hl\" (UniqueName: \"kubernetes.io/projected/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-kube-api-access-kq5hl\") pod \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\" (UID: \"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8\") " Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.156465 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8" (UID: "16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.161376 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-kube-api-access-kq5hl" (OuterVolumeSpecName: "kube-api-access-kq5hl") pod "16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8" (UID: "16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8"). InnerVolumeSpecName "kube-api-access-kq5hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.258008 4690 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.258304 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq5hl\" (UniqueName: \"kubernetes.io/projected/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8-kube-api-access-kq5hl\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.689810 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerStarted","Data":"4cade190ea7f13ff5eb126bfc603dd58763302b591bdc98bcb699576527bb663"} Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.692047 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" event={"ID":"16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8","Type":"ContainerDied","Data":"523043849ff1f4c50e29decc406ed18ed8b4abb0964daa34bd079b455ba7932f"} Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.692091 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="523043849ff1f4c50e29decc406ed18ed8b4abb0964daa34bd079b455ba7932f" Dec 03 13:33:43 crc kubenswrapper[4690]: I1203 13:33:43.692100 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-de32-account-create-update-d5v2x" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.140080 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176023 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hvqqp"] Dec 03 13:33:44 crc kubenswrapper[4690]: E1203 13:33:44.176529 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8137b9f0-e7c6-4134-ae61-5599982fa141" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176552 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8137b9f0-e7c6-4134-ae61-5599982fa141" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: E1203 13:33:44.176565 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2478ef86-8c11-4d36-88b1-eea0a3b2310a" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176572 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="2478ef86-8c11-4d36-88b1-eea0a3b2310a" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: E1203 13:33:44.176591 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176597 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: E1203 13:33:44.176616 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad6e77ea-b90d-4def-8ab1-d761fc41ebe3" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176622 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad6e77ea-b90d-4def-8ab1-d761fc41ebe3" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: E1203 13:33:44.176638 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf7fd9d-669f-4d15-a432-8b796d5f1de5" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176644 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf7fd9d-669f-4d15-a432-8b796d5f1de5" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: E1203 13:33:44.176663 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbddf981-d575-4517-b0e6-f494b9a92fe6" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176669 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbddf981-d575-4517-b0e6-f494b9a92fe6" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176886 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad6e77ea-b90d-4def-8ab1-d761fc41ebe3" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176904 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bf7fd9d-669f-4d15-a432-8b796d5f1de5" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176916 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="2478ef86-8c11-4d36-88b1-eea0a3b2310a" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176929 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="8137b9f0-e7c6-4134-ae61-5599982fa141" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176936 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8" containerName="mariadb-account-create-update" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.176944 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbddf981-d575-4517-b0e6-f494b9a92fe6" containerName="mariadb-database-create" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.177700 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.181162 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.181383 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-l58jm" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.184054 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.204513 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hvqqp"] Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.283277 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-config-data\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.283331 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-scripts\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.283411 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76v4z\" (UniqueName: \"kubernetes.io/projected/43129653-d1cd-4ced-b120-33be343c3d9d-kube-api-access-76v4z\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.283436 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.385734 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76v4z\" (UniqueName: \"kubernetes.io/projected/43129653-d1cd-4ced-b120-33be343c3d9d-kube-api-access-76v4z\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.385780 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.385955 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-config-data\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.385982 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-scripts\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.391757 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-scripts\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.392161 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-config-data\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.401641 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.412196 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76v4z\" (UniqueName: \"kubernetes.io/projected/43129653-d1cd-4ced-b120-33be343c3d9d-kube-api-access-76v4z\") pod \"nova-cell0-conductor-db-sync-hvqqp\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:44 crc kubenswrapper[4690]: I1203 13:33:44.494048 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:33:45 crc kubenswrapper[4690]: I1203 13:33:45.016250 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hvqqp"] Dec 03 13:33:45 crc kubenswrapper[4690]: I1203 13:33:45.720113 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" event={"ID":"43129653-d1cd-4ced-b120-33be343c3d9d","Type":"ContainerStarted","Data":"619b74c0587fed892cf2a7148e4b2e50c5604961ceba5724e4b18d004bcb8770"} Dec 03 13:33:45 crc kubenswrapper[4690]: I1203 13:33:45.720184 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:33:45 crc kubenswrapper[4690]: I1203 13:33:45.749375 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.589291234 podStartE2EDuration="13.749326199s" podCreationTimestamp="2025-12-03 13:33:32 +0000 UTC" firstStartedPulling="2025-12-03 13:33:33.286943208 +0000 UTC m=+1459.267863641" lastFinishedPulling="2025-12-03 13:33:42.446978173 +0000 UTC m=+1468.427898606" observedRunningTime="2025-12-03 13:33:45.742118652 +0000 UTC m=+1471.723039105" watchObservedRunningTime="2025-12-03 13:33:45.749326199 +0000 UTC m=+1471.730246642" Dec 03 13:33:46 crc kubenswrapper[4690]: I1203 13:33:46.824200 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:33:46 crc kubenswrapper[4690]: I1203 13:33:46.824672 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:33:46 crc kubenswrapper[4690]: I1203 13:33:46.824747 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:33:46 crc kubenswrapper[4690]: I1203 13:33:46.826131 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d13f0eb20e5e81edaffaa899e1251f34d5ba5b9191f18b642a8f24f971ecc07d"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:33:46 crc kubenswrapper[4690]: I1203 13:33:46.826208 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://d13f0eb20e5e81edaffaa899e1251f34d5ba5b9191f18b642a8f24f971ecc07d" gracePeriod=600 Dec 03 13:33:47 crc kubenswrapper[4690]: I1203 13:33:47.743393 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="d13f0eb20e5e81edaffaa899e1251f34d5ba5b9191f18b642a8f24f971ecc07d" exitCode=0 Dec 03 13:33:47 crc kubenswrapper[4690]: I1203 13:33:47.743462 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"d13f0eb20e5e81edaffaa899e1251f34d5ba5b9191f18b642a8f24f971ecc07d"} Dec 03 13:33:47 crc kubenswrapper[4690]: I1203 13:33:47.743807 4690 scope.go:117] "RemoveContainer" containerID="22570d41f66f8cf71e73ec062157689b8763204a45407561d1852ebc8cede0d1" Dec 03 13:33:49 crc kubenswrapper[4690]: I1203 13:33:49.772351 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2"} Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.091600 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.092156 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="ceilometer-central-agent" containerID="cri-o://3e52b546c1b666c3e15e4a412825dee0bc44d343e3dbc9732b55d0542bac5216" gracePeriod=30 Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.092474 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="ceilometer-notification-agent" containerID="cri-o://823c62f78dfd2ed299db0b1cd0ac3f4f75381387f193254c72c3ced9000479e1" gracePeriod=30 Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.092508 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="sg-core" containerID="cri-o://760ccb4bdd5df2a779c960b37cd6966a39dcabae942397164f21d5347e8b3f3c" gracePeriod=30 Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.092430 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="proxy-httpd" containerID="cri-o://4cade190ea7f13ff5eb126bfc603dd58763302b591bdc98bcb699576527bb663" gracePeriod=30 Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.804769 4690 generic.go:334] "Generic (PLEG): container finished" podID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerID="4cade190ea7f13ff5eb126bfc603dd58763302b591bdc98bcb699576527bb663" exitCode=0 Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.806079 4690 generic.go:334] "Generic (PLEG): container finished" podID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerID="760ccb4bdd5df2a779c960b37cd6966a39dcabae942397164f21d5347e8b3f3c" exitCode=2 Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.804910 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerDied","Data":"4cade190ea7f13ff5eb126bfc603dd58763302b591bdc98bcb699576527bb663"} Dec 03 13:33:51 crc kubenswrapper[4690]: I1203 13:33:51.806164 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerDied","Data":"760ccb4bdd5df2a779c960b37cd6966a39dcabae942397164f21d5347e8b3f3c"} Dec 03 13:33:54 crc kubenswrapper[4690]: I1203 13:33:54.842711 4690 generic.go:334] "Generic (PLEG): container finished" podID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerID="823c62f78dfd2ed299db0b1cd0ac3f4f75381387f193254c72c3ced9000479e1" exitCode=0 Dec 03 13:33:54 crc kubenswrapper[4690]: I1203 13:33:54.844062 4690 generic.go:334] "Generic (PLEG): container finished" podID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerID="3e52b546c1b666c3e15e4a412825dee0bc44d343e3dbc9732b55d0542bac5216" exitCode=0 Dec 03 13:33:54 crc kubenswrapper[4690]: I1203 13:33:54.842877 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerDied","Data":"823c62f78dfd2ed299db0b1cd0ac3f4f75381387f193254c72c3ced9000479e1"} Dec 03 13:33:54 crc kubenswrapper[4690]: I1203 13:33:54.844218 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerDied","Data":"3e52b546c1b666c3e15e4a412825dee0bc44d343e3dbc9732b55d0542bac5216"} Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.846888 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.904034 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" event={"ID":"43129653-d1cd-4ced-b120-33be343c3d9d","Type":"ContainerStarted","Data":"41fa04cdec6857177bb3fc18b10b5e4e11a9beade25ddd077b1f275dc2a49aa8"} Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.911291 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87b18c03-5cae-4ee6-9634-f3d832d57d66","Type":"ContainerDied","Data":"bb858caaf6787b0052a61b977cc1e28fc7464859fc635052df99513767ca2431"} Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.911369 4690 scope.go:117] "RemoveContainer" containerID="4cade190ea7f13ff5eb126bfc603dd58763302b591bdc98bcb699576527bb663" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.911458 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.937908 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" podStartSLOduration=1.439540822 podStartE2EDuration="13.937875553s" podCreationTimestamp="2025-12-03 13:33:44 +0000 UTC" firstStartedPulling="2025-12-03 13:33:45.015178819 +0000 UTC m=+1470.996099252" lastFinishedPulling="2025-12-03 13:33:57.51351355 +0000 UTC m=+1483.494433983" observedRunningTime="2025-12-03 13:33:57.936843466 +0000 UTC m=+1483.917763899" watchObservedRunningTime="2025-12-03 13:33:57.937875553 +0000 UTC m=+1483.918795986" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.944167 4690 scope.go:117] "RemoveContainer" containerID="760ccb4bdd5df2a779c960b37cd6966a39dcabae942397164f21d5347e8b3f3c" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.957117 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-config-data\") pod \"87b18c03-5cae-4ee6-9634-f3d832d57d66\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.957402 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-combined-ca-bundle\") pod \"87b18c03-5cae-4ee6-9634-f3d832d57d66\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.957488 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-scripts\") pod \"87b18c03-5cae-4ee6-9634-f3d832d57d66\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.957535 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdd4c\" (UniqueName: \"kubernetes.io/projected/87b18c03-5cae-4ee6-9634-f3d832d57d66-kube-api-access-mdd4c\") pod \"87b18c03-5cae-4ee6-9634-f3d832d57d66\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.957616 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-sg-core-conf-yaml\") pod \"87b18c03-5cae-4ee6-9634-f3d832d57d66\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.957655 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-run-httpd\") pod \"87b18c03-5cae-4ee6-9634-f3d832d57d66\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.957711 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-log-httpd\") pod \"87b18c03-5cae-4ee6-9634-f3d832d57d66\" (UID: \"87b18c03-5cae-4ee6-9634-f3d832d57d66\") " Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.959671 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "87b18c03-5cae-4ee6-9634-f3d832d57d66" (UID: "87b18c03-5cae-4ee6-9634-f3d832d57d66"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.961148 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "87b18c03-5cae-4ee6-9634-f3d832d57d66" (UID: "87b18c03-5cae-4ee6-9634-f3d832d57d66"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.974507 4690 scope.go:117] "RemoveContainer" containerID="823c62f78dfd2ed299db0b1cd0ac3f4f75381387f193254c72c3ced9000479e1" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.979076 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-scripts" (OuterVolumeSpecName: "scripts") pod "87b18c03-5cae-4ee6-9634-f3d832d57d66" (UID: "87b18c03-5cae-4ee6-9634-f3d832d57d66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.990154 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87b18c03-5cae-4ee6-9634-f3d832d57d66-kube-api-access-mdd4c" (OuterVolumeSpecName: "kube-api-access-mdd4c") pod "87b18c03-5cae-4ee6-9634-f3d832d57d66" (UID: "87b18c03-5cae-4ee6-9634-f3d832d57d66"). InnerVolumeSpecName "kube-api-access-mdd4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:33:57 crc kubenswrapper[4690]: I1203 13:33:57.999619 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "87b18c03-5cae-4ee6-9634-f3d832d57d66" (UID: "87b18c03-5cae-4ee6-9634-f3d832d57d66"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.061229 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.061271 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdd4c\" (UniqueName: \"kubernetes.io/projected/87b18c03-5cae-4ee6-9634-f3d832d57d66-kube-api-access-mdd4c\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.061287 4690 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.061302 4690 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.061315 4690 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87b18c03-5cae-4ee6-9634-f3d832d57d66-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.071076 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87b18c03-5cae-4ee6-9634-f3d832d57d66" (UID: "87b18c03-5cae-4ee6-9634-f3d832d57d66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.097289 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-config-data" (OuterVolumeSpecName: "config-data") pod "87b18c03-5cae-4ee6-9634-f3d832d57d66" (UID: "87b18c03-5cae-4ee6-9634-f3d832d57d66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.126145 4690 scope.go:117] "RemoveContainer" containerID="3e52b546c1b666c3e15e4a412825dee0bc44d343e3dbc9732b55d0542bac5216" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.163261 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.163305 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b18c03-5cae-4ee6-9634-f3d832d57d66-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.252138 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.261525 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295097 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:58 crc kubenswrapper[4690]: E1203 13:33:58.295533 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="ceilometer-notification-agent" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295553 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="ceilometer-notification-agent" Dec 03 13:33:58 crc kubenswrapper[4690]: E1203 13:33:58.295572 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="proxy-httpd" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295579 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="proxy-httpd" Dec 03 13:33:58 crc kubenswrapper[4690]: E1203 13:33:58.295603 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="ceilometer-central-agent" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295610 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="ceilometer-central-agent" Dec 03 13:33:58 crc kubenswrapper[4690]: E1203 13:33:58.295626 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="sg-core" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295631 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="sg-core" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295815 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="sg-core" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295830 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="ceilometer-notification-agent" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295843 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="proxy-httpd" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.295853 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" containerName="ceilometer-central-agent" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.299681 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.303451 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.303461 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.332016 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87b18c03-5cae-4ee6-9634-f3d832d57d66" path="/var/lib/kubelet/pods/87b18c03-5cae-4ee6-9634-f3d832d57d66/volumes" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.332996 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.367318 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-scripts\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.367367 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-run-httpd\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.367397 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-log-httpd\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.367518 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.367539 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.367565 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-config-data\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.367669 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgjlg\" (UniqueName: \"kubernetes.io/projected/a9d00402-b672-47aa-94ec-caf3badd5e42-kube-api-access-hgjlg\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.469759 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgjlg\" (UniqueName: \"kubernetes.io/projected/a9d00402-b672-47aa-94ec-caf3badd5e42-kube-api-access-hgjlg\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.469884 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-scripts\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.469925 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-run-httpd\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.469962 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-log-httpd\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.470049 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.470075 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.470502 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-run-httpd\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.470656 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-log-httpd\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.470922 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-config-data\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.476617 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.477566 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-scripts\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.477916 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.478954 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-config-data\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.489599 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgjlg\" (UniqueName: \"kubernetes.io/projected/a9d00402-b672-47aa-94ec-caf3badd5e42-kube-api-access-hgjlg\") pod \"ceilometer-0\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " pod="openstack/ceilometer-0" Dec 03 13:33:58 crc kubenswrapper[4690]: I1203 13:33:58.628516 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:33:59 crc kubenswrapper[4690]: I1203 13:33:59.344009 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:33:59 crc kubenswrapper[4690]: I1203 13:33:59.996935 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerStarted","Data":"4e4e1a4af688e3f591666beda2adcc6bbadd85a6c30a8a2046b6fda6c4a398b3"} Dec 03 13:34:01 crc kubenswrapper[4690]: I1203 13:34:01.007444 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerStarted","Data":"1626208b68acaf5420f6308e51ce5ee22868aed291e50beb0726869617e8a292"} Dec 03 13:34:01 crc kubenswrapper[4690]: I1203 13:34:01.040640 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:02 crc kubenswrapper[4690]: I1203 13:34:02.021710 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerStarted","Data":"815a323f712be6e75a4b88a905c1e3d8a05ef70018c179bebb9186b0c7cf0a1a"} Dec 03 13:34:08 crc kubenswrapper[4690]: I1203 13:34:08.095481 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerStarted","Data":"f785e9f133be17054934c08681584c2d2769c1dffe87ee9725b21a7fe3d75d67"} Dec 03 13:34:10 crc kubenswrapper[4690]: I1203 13:34:10.120975 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerStarted","Data":"be48e9957d08bf74f5fa9c1dc6ab4cfe4cb8f6845731ce4df922ecd80d1b4b71"} Dec 03 13:34:10 crc kubenswrapper[4690]: I1203 13:34:10.121192 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="ceilometer-central-agent" containerID="cri-o://1626208b68acaf5420f6308e51ce5ee22868aed291e50beb0726869617e8a292" gracePeriod=30 Dec 03 13:34:10 crc kubenswrapper[4690]: I1203 13:34:10.121237 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="proxy-httpd" containerID="cri-o://be48e9957d08bf74f5fa9c1dc6ab4cfe4cb8f6845731ce4df922ecd80d1b4b71" gracePeriod=30 Dec 03 13:34:10 crc kubenswrapper[4690]: I1203 13:34:10.121299 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="ceilometer-notification-agent" containerID="cri-o://815a323f712be6e75a4b88a905c1e3d8a05ef70018c179bebb9186b0c7cf0a1a" gracePeriod=30 Dec 03 13:34:10 crc kubenswrapper[4690]: I1203 13:34:10.121346 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="sg-core" containerID="cri-o://f785e9f133be17054934c08681584c2d2769c1dffe87ee9725b21a7fe3d75d67" gracePeriod=30 Dec 03 13:34:10 crc kubenswrapper[4690]: I1203 13:34:10.122834 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:34:10 crc kubenswrapper[4690]: I1203 13:34:10.152610 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5391980050000003 podStartE2EDuration="12.15259041s" podCreationTimestamp="2025-12-03 13:33:58 +0000 UTC" firstStartedPulling="2025-12-03 13:33:59.355596144 +0000 UTC m=+1485.336516577" lastFinishedPulling="2025-12-03 13:34:08.968988549 +0000 UTC m=+1494.949908982" observedRunningTime="2025-12-03 13:34:10.14650949 +0000 UTC m=+1496.127429923" watchObservedRunningTime="2025-12-03 13:34:10.15259041 +0000 UTC m=+1496.133510833" Dec 03 13:34:11 crc kubenswrapper[4690]: I1203 13:34:11.134022 4690 generic.go:334] "Generic (PLEG): container finished" podID="3ca321dd-bd9d-4039-aa80-eb386d63dcb4" containerID="eb7088271fec7391af00bc7bb58cc71b4f9de4efd677b58db0288059fc39f036" exitCode=0 Dec 03 13:34:11 crc kubenswrapper[4690]: I1203 13:34:11.134100 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2xpmv" event={"ID":"3ca321dd-bd9d-4039-aa80-eb386d63dcb4","Type":"ContainerDied","Data":"eb7088271fec7391af00bc7bb58cc71b4f9de4efd677b58db0288059fc39f036"} Dec 03 13:34:11 crc kubenswrapper[4690]: I1203 13:34:11.137954 4690 generic.go:334] "Generic (PLEG): container finished" podID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerID="be48e9957d08bf74f5fa9c1dc6ab4cfe4cb8f6845731ce4df922ecd80d1b4b71" exitCode=0 Dec 03 13:34:11 crc kubenswrapper[4690]: I1203 13:34:11.138000 4690 generic.go:334] "Generic (PLEG): container finished" podID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerID="f785e9f133be17054934c08681584c2d2769c1dffe87ee9725b21a7fe3d75d67" exitCode=2 Dec 03 13:34:11 crc kubenswrapper[4690]: I1203 13:34:11.138013 4690 generic.go:334] "Generic (PLEG): container finished" podID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerID="815a323f712be6e75a4b88a905c1e3d8a05ef70018c179bebb9186b0c7cf0a1a" exitCode=0 Dec 03 13:34:11 crc kubenswrapper[4690]: I1203 13:34:11.138020 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerDied","Data":"be48e9957d08bf74f5fa9c1dc6ab4cfe4cb8f6845731ce4df922ecd80d1b4b71"} Dec 03 13:34:11 crc kubenswrapper[4690]: I1203 13:34:11.138092 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerDied","Data":"f785e9f133be17054934c08681584c2d2769c1dffe87ee9725b21a7fe3d75d67"} Dec 03 13:34:11 crc kubenswrapper[4690]: I1203 13:34:11.138109 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerDied","Data":"815a323f712be6e75a4b88a905c1e3d8a05ef70018c179bebb9186b0c7cf0a1a"} Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.613127 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.809252 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-config\") pod \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.809359 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-combined-ca-bundle\") pod \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.809387 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsfr8\" (UniqueName: \"kubernetes.io/projected/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-kube-api-access-lsfr8\") pod \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\" (UID: \"3ca321dd-bd9d-4039-aa80-eb386d63dcb4\") " Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.816116 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-kube-api-access-lsfr8" (OuterVolumeSpecName: "kube-api-access-lsfr8") pod "3ca321dd-bd9d-4039-aa80-eb386d63dcb4" (UID: "3ca321dd-bd9d-4039-aa80-eb386d63dcb4"). InnerVolumeSpecName "kube-api-access-lsfr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.842568 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-config" (OuterVolumeSpecName: "config") pod "3ca321dd-bd9d-4039-aa80-eb386d63dcb4" (UID: "3ca321dd-bd9d-4039-aa80-eb386d63dcb4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.847025 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ca321dd-bd9d-4039-aa80-eb386d63dcb4" (UID: "3ca321dd-bd9d-4039-aa80-eb386d63dcb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.912249 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.912288 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:12 crc kubenswrapper[4690]: I1203 13:34:12.912300 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsfr8\" (UniqueName: \"kubernetes.io/projected/3ca321dd-bd9d-4039-aa80-eb386d63dcb4-kube-api-access-lsfr8\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.160620 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2xpmv" event={"ID":"3ca321dd-bd9d-4039-aa80-eb386d63dcb4","Type":"ContainerDied","Data":"8fdeff4385a0edc212a40024bd94720d185c7facc4fd42829603ba46d6d7b9fb"} Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.160666 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fdeff4385a0edc212a40024bd94720d185c7facc4fd42829603ba46d6d7b9fb" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.160674 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2xpmv" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.333994 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-sw2bh"] Dec 03 13:34:13 crc kubenswrapper[4690]: E1203 13:34:13.337611 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca321dd-bd9d-4039-aa80-eb386d63dcb4" containerName="neutron-db-sync" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.337659 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca321dd-bd9d-4039-aa80-eb386d63dcb4" containerName="neutron-db-sync" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.337913 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca321dd-bd9d-4039-aa80-eb386d63dcb4" containerName="neutron-db-sync" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.342481 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.369058 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-sw2bh"] Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.423828 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-65497df674-vnvht"] Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.426307 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.429374 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.429855 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.429919 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.434221 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-q5ggj" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.453814 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65497df674-vnvht"] Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.527082 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-config\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.529733 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-config\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.529809 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b45q6\" (UniqueName: \"kubernetes.io/projected/e70496e1-182e-4071-9170-ab8bf54a14ee-kube-api-access-b45q6\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.529920 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.530031 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-httpd-config\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.530055 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgnn9\" (UniqueName: \"kubernetes.io/projected/5ce567ee-468c-4f48-835d-a09de85f5d93-kube-api-access-vgnn9\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.530176 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.530252 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-combined-ca-bundle\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.530329 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.530355 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-ovndb-tls-certs\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.530458 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635122 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-config\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635177 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-config\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635199 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b45q6\" (UniqueName: \"kubernetes.io/projected/e70496e1-182e-4071-9170-ab8bf54a14ee-kube-api-access-b45q6\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635230 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635263 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-httpd-config\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635282 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgnn9\" (UniqueName: \"kubernetes.io/projected/5ce567ee-468c-4f48-835d-a09de85f5d93-kube-api-access-vgnn9\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635320 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635344 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-combined-ca-bundle\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635367 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-ovndb-tls-certs\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635385 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.635415 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.636400 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.636998 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.637299 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-config\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.639498 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.642896 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-httpd-config\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.643311 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-config\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.656529 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.657661 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-ovndb-tls-certs\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.666360 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-combined-ca-bundle\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.679636 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgnn9\" (UniqueName: \"kubernetes.io/projected/5ce567ee-468c-4f48-835d-a09de85f5d93-kube-api-access-vgnn9\") pod \"neutron-65497df674-vnvht\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.680693 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b45q6\" (UniqueName: \"kubernetes.io/projected/e70496e1-182e-4071-9170-ab8bf54a14ee-kube-api-access-b45q6\") pod \"dnsmasq-dns-5c9776ccc5-sw2bh\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.745581 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:13 crc kubenswrapper[4690]: I1203 13:34:13.975150 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:14 crc kubenswrapper[4690]: I1203 13:34:14.461200 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65497df674-vnvht"] Dec 03 13:34:14 crc kubenswrapper[4690]: I1203 13:34:14.676843 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-sw2bh"] Dec 03 13:34:14 crc kubenswrapper[4690]: W1203 13:34:14.685084 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode70496e1_182e_4071_9170_ab8bf54a14ee.slice/crio-4261caa0df63055f253562529221cbcf76e2264af0c1f34abf171cdeff076f6b WatchSource:0}: Error finding container 4261caa0df63055f253562529221cbcf76e2264af0c1f34abf171cdeff076f6b: Status 404 returned error can't find the container with id 4261caa0df63055f253562529221cbcf76e2264af0c1f34abf171cdeff076f6b Dec 03 13:34:15 crc kubenswrapper[4690]: I1203 13:34:15.200254 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65497df674-vnvht" event={"ID":"5ce567ee-468c-4f48-835d-a09de85f5d93","Type":"ContainerStarted","Data":"d97a168ca3980ef383c6447c5f7c511c70cd172a3752a6d3089e3d5599bf82f5"} Dec 03 13:34:15 crc kubenswrapper[4690]: I1203 13:34:15.202489 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" event={"ID":"e70496e1-182e-4071-9170-ab8bf54a14ee","Type":"ContainerStarted","Data":"4261caa0df63055f253562529221cbcf76e2264af0c1f34abf171cdeff076f6b"} Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.215362 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65497df674-vnvht" event={"ID":"5ce567ee-468c-4f48-835d-a09de85f5d93","Type":"ContainerStarted","Data":"be72616885f72bda5f6b02089fdc438a90dee30aa3513d38e8768270f11c787f"} Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.217506 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" event={"ID":"e70496e1-182e-4071-9170-ab8bf54a14ee","Type":"ContainerStarted","Data":"c9385cdab24a1523c9f7967aaccad46c7f31abf218fc1aff032867ebf7cf3bde"} Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.609386 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d87869f9-cmzfc"] Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.624543 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.628546 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.628805 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.645008 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d87869f9-cmzfc"] Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.739807 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-combined-ca-bundle\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.740009 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-httpd-config\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.740082 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-config\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.740262 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-ovndb-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.740354 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-internal-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.740501 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-public-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.740547 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj2c8\" (UniqueName: \"kubernetes.io/projected/7d486acd-1716-4926-be13-3e67b3106c02-kube-api-access-dj2c8\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.842932 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-httpd-config\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.843076 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-config\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.843217 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-ovndb-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.843295 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-internal-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.843336 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-public-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.844116 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj2c8\" (UniqueName: \"kubernetes.io/projected/7d486acd-1716-4926-be13-3e67b3106c02-kube-api-access-dj2c8\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.844205 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-combined-ca-bundle\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.851395 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-combined-ca-bundle\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.852607 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-httpd-config\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.853360 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-ovndb-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.853630 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-config\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.855045 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-public-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.856669 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d486acd-1716-4926-be13-3e67b3106c02-internal-tls-certs\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:16 crc kubenswrapper[4690]: I1203 13:34:16.870854 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj2c8\" (UniqueName: \"kubernetes.io/projected/7d486acd-1716-4926-be13-3e67b3106c02-kube-api-access-dj2c8\") pod \"neutron-d87869f9-cmzfc\" (UID: \"7d486acd-1716-4926-be13-3e67b3106c02\") " pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:17 crc kubenswrapper[4690]: I1203 13:34:17.037002 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:17 crc kubenswrapper[4690]: I1203 13:34:17.244274 4690 generic.go:334] "Generic (PLEG): container finished" podID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerID="1626208b68acaf5420f6308e51ce5ee22868aed291e50beb0726869617e8a292" exitCode=0 Dec 03 13:34:17 crc kubenswrapper[4690]: I1203 13:34:17.244922 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerDied","Data":"1626208b68acaf5420f6308e51ce5ee22868aed291e50beb0726869617e8a292"} Dec 03 13:34:17 crc kubenswrapper[4690]: I1203 13:34:17.256073 4690 generic.go:334] "Generic (PLEG): container finished" podID="e70496e1-182e-4071-9170-ab8bf54a14ee" containerID="c9385cdab24a1523c9f7967aaccad46c7f31abf218fc1aff032867ebf7cf3bde" exitCode=0 Dec 03 13:34:17 crc kubenswrapper[4690]: I1203 13:34:17.256134 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" event={"ID":"e70496e1-182e-4071-9170-ab8bf54a14ee","Type":"ContainerDied","Data":"c9385cdab24a1523c9f7967aaccad46c7f31abf218fc1aff032867ebf7cf3bde"} Dec 03 13:34:17 crc kubenswrapper[4690]: I1203 13:34:17.686625 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d87869f9-cmzfc"] Dec 03 13:34:17 crc kubenswrapper[4690]: I1203 13:34:17.901753 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.079352 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-scripts\") pod \"a9d00402-b672-47aa-94ec-caf3badd5e42\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.079413 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgjlg\" (UniqueName: \"kubernetes.io/projected/a9d00402-b672-47aa-94ec-caf3badd5e42-kube-api-access-hgjlg\") pod \"a9d00402-b672-47aa-94ec-caf3badd5e42\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.079481 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-run-httpd\") pod \"a9d00402-b672-47aa-94ec-caf3badd5e42\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.080013 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a9d00402-b672-47aa-94ec-caf3badd5e42" (UID: "a9d00402-b672-47aa-94ec-caf3badd5e42"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.080466 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-config-data\") pod \"a9d00402-b672-47aa-94ec-caf3badd5e42\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.080599 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-log-httpd\") pod \"a9d00402-b672-47aa-94ec-caf3badd5e42\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.080741 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-sg-core-conf-yaml\") pod \"a9d00402-b672-47aa-94ec-caf3badd5e42\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.080782 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-combined-ca-bundle\") pod \"a9d00402-b672-47aa-94ec-caf3badd5e42\" (UID: \"a9d00402-b672-47aa-94ec-caf3badd5e42\") " Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.081144 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a9d00402-b672-47aa-94ec-caf3badd5e42" (UID: "a9d00402-b672-47aa-94ec-caf3badd5e42"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.081623 4690 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.081641 4690 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9d00402-b672-47aa-94ec-caf3badd5e42-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.087194 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-scripts" (OuterVolumeSpecName: "scripts") pod "a9d00402-b672-47aa-94ec-caf3badd5e42" (UID: "a9d00402-b672-47aa-94ec-caf3badd5e42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.099625 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d00402-b672-47aa-94ec-caf3badd5e42-kube-api-access-hgjlg" (OuterVolumeSpecName: "kube-api-access-hgjlg") pod "a9d00402-b672-47aa-94ec-caf3badd5e42" (UID: "a9d00402-b672-47aa-94ec-caf3badd5e42"). InnerVolumeSpecName "kube-api-access-hgjlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.182624 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a9d00402-b672-47aa-94ec-caf3badd5e42" (UID: "a9d00402-b672-47aa-94ec-caf3badd5e42"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.184603 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.184634 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgjlg\" (UniqueName: \"kubernetes.io/projected/a9d00402-b672-47aa-94ec-caf3badd5e42-kube-api-access-hgjlg\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.184646 4690 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.253581 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-config-data" (OuterVolumeSpecName: "config-data") pod "a9d00402-b672-47aa-94ec-caf3badd5e42" (UID: "a9d00402-b672-47aa-94ec-caf3badd5e42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.265599 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9d00402-b672-47aa-94ec-caf3badd5e42" (UID: "a9d00402-b672-47aa-94ec-caf3badd5e42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.272042 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9d00402-b672-47aa-94ec-caf3badd5e42","Type":"ContainerDied","Data":"4e4e1a4af688e3f591666beda2adcc6bbadd85a6c30a8a2046b6fda6c4a398b3"} Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.272109 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.272117 4690 scope.go:117] "RemoveContainer" containerID="be48e9957d08bf74f5fa9c1dc6ab4cfe4cb8f6845731ce4df922ecd80d1b4b71" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.281580 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d87869f9-cmzfc" event={"ID":"7d486acd-1716-4926-be13-3e67b3106c02","Type":"ContainerStarted","Data":"89186e9dce85488480824653f2be948420d92ee44cd6c10d3e4e1b159335be2a"} Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.286315 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.286351 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9d00402-b672-47aa-94ec-caf3badd5e42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.287421 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65497df674-vnvht" event={"ID":"5ce567ee-468c-4f48-835d-a09de85f5d93","Type":"ContainerStarted","Data":"a5717985d3761a304110d932fdcd5d6ba72e5ce8b97174b2086f03ef0883d685"} Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.299820 4690 scope.go:117] "RemoveContainer" containerID="f785e9f133be17054934c08681584c2d2769c1dffe87ee9725b21a7fe3d75d67" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.368651 4690 scope.go:117] "RemoveContainer" containerID="815a323f712be6e75a4b88a905c1e3d8a05ef70018c179bebb9186b0c7cf0a1a" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.391466 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.400903 4690 scope.go:117] "RemoveContainer" containerID="1626208b68acaf5420f6308e51ce5ee22868aed291e50beb0726869617e8a292" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.416954 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.425859 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:18 crc kubenswrapper[4690]: E1203 13:34:18.426671 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="proxy-httpd" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.426699 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="proxy-httpd" Dec 03 13:34:18 crc kubenswrapper[4690]: E1203 13:34:18.426738 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="ceilometer-notification-agent" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.426748 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="ceilometer-notification-agent" Dec 03 13:34:18 crc kubenswrapper[4690]: E1203 13:34:18.426768 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="ceilometer-central-agent" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.426777 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="ceilometer-central-agent" Dec 03 13:34:18 crc kubenswrapper[4690]: E1203 13:34:18.426786 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="sg-core" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.426794 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="sg-core" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.427117 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="proxy-httpd" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.427147 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="ceilometer-notification-agent" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.427158 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="sg-core" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.427177 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" containerName="ceilometer-central-agent" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.430439 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.434182 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.445856 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.446174 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.593626 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llcrf\" (UniqueName: \"kubernetes.io/projected/01b61175-58d3-4291-9b9f-c791281fe495-kube-api-access-llcrf\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.593696 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-scripts\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.593751 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.593777 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-log-httpd\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.593819 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.593863 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-run-httpd\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.593908 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-config-data\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.695896 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.696270 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-log-httpd\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.696383 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.696494 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-run-httpd\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.696598 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-config-data\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.696747 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llcrf\" (UniqueName: \"kubernetes.io/projected/01b61175-58d3-4291-9b9f-c791281fe495-kube-api-access-llcrf\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.696845 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-scripts\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.697090 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-run-httpd\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.697349 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-log-httpd\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.703081 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.703326 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-config-data\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.704028 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-scripts\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.709315 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.719390 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llcrf\" (UniqueName: \"kubernetes.io/projected/01b61175-58d3-4291-9b9f-c791281fe495-kube-api-access-llcrf\") pod \"ceilometer-0\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " pod="openstack/ceilometer-0" Dec 03 13:34:18 crc kubenswrapper[4690]: I1203 13:34:18.817320 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.302544 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" event={"ID":"e70496e1-182e-4071-9170-ab8bf54a14ee","Type":"ContainerStarted","Data":"192f75d3c726653dccdb15141796efc6d4b6c451829082723fd8d74f18e4583b"} Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.303214 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.343024 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-65497df674-vnvht" podStartSLOduration=6.343004923 podStartE2EDuration="6.343004923s" podCreationTimestamp="2025-12-03 13:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:34:19.335770052 +0000 UTC m=+1505.316690495" watchObservedRunningTime="2025-12-03 13:34:19.343004923 +0000 UTC m=+1505.323925356" Dec 03 13:34:19 crc kubenswrapper[4690]: W1203 13:34:19.381174 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01b61175_58d3_4291_9b9f_c791281fe495.slice/crio-0120fa55e3fdb606c0b15ec769e06b0aa7854fd476de2513e57e2e74bfec82ae WatchSource:0}: Error finding container 0120fa55e3fdb606c0b15ec769e06b0aa7854fd476de2513e57e2e74bfec82ae: Status 404 returned error can't find the container with id 0120fa55e3fdb606c0b15ec769e06b0aa7854fd476de2513e57e2e74bfec82ae Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.381699 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.761050 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l5z6x"] Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.763589 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.789097 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l5z6x"] Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.930291 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-utilities\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.930410 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-catalog-content\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:19 crc kubenswrapper[4690]: I1203 13:34:19.930449 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blsxv\" (UniqueName: \"kubernetes.io/projected/7d64a510-19ae-40a6-bf85-e538894ef917-kube-api-access-blsxv\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.033618 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-utilities\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.034217 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-catalog-content\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.034284 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blsxv\" (UniqueName: \"kubernetes.io/projected/7d64a510-19ae-40a6-bf85-e538894ef917-kube-api-access-blsxv\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.034541 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-utilities\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.035381 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-catalog-content\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.058785 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blsxv\" (UniqueName: \"kubernetes.io/projected/7d64a510-19ae-40a6-bf85-e538894ef917-kube-api-access-blsxv\") pod \"community-operators-l5z6x\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.089504 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.346968 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9d00402-b672-47aa-94ec-caf3badd5e42" path="/var/lib/kubelet/pods/a9d00402-b672-47aa-94ec-caf3badd5e42/volumes" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.349688 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerStarted","Data":"7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da"} Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.349714 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerStarted","Data":"0120fa55e3fdb606c0b15ec769e06b0aa7854fd476de2513e57e2e74bfec82ae"} Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.349731 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.349742 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d87869f9-cmzfc" event={"ID":"7d486acd-1716-4926-be13-3e67b3106c02","Type":"ContainerStarted","Data":"2d6b6b2e1810451f6ac2d1d1da80479c8ebfb9db2067c967586e78d5737f7cf6"} Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.349752 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.349761 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d87869f9-cmzfc" event={"ID":"7d486acd-1716-4926-be13-3e67b3106c02","Type":"ContainerStarted","Data":"b84c162cd2ee97cff80ea8e815c44a6b51621d3828ee319e58d3f32a7e0e4d5f"} Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.398643 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" podStartSLOduration=7.398617538 podStartE2EDuration="7.398617538s" podCreationTimestamp="2025-12-03 13:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:34:20.382104172 +0000 UTC m=+1506.363024625" watchObservedRunningTime="2025-12-03 13:34:20.398617538 +0000 UTC m=+1506.379537971" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.445478 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d87869f9-cmzfc" podStartSLOduration=4.445449603 podStartE2EDuration="4.445449603s" podCreationTimestamp="2025-12-03 13:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:34:20.417907927 +0000 UTC m=+1506.398828380" watchObservedRunningTime="2025-12-03 13:34:20.445449603 +0000 UTC m=+1506.426370036" Dec 03 13:34:20 crc kubenswrapper[4690]: I1203 13:34:20.689710 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l5z6x"] Dec 03 13:34:21 crc kubenswrapper[4690]: I1203 13:34:21.385525 4690 generic.go:334] "Generic (PLEG): container finished" podID="7d64a510-19ae-40a6-bf85-e538894ef917" containerID="40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577" exitCode=0 Dec 03 13:34:21 crc kubenswrapper[4690]: I1203 13:34:21.386316 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l5z6x" event={"ID":"7d64a510-19ae-40a6-bf85-e538894ef917","Type":"ContainerDied","Data":"40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577"} Dec 03 13:34:21 crc kubenswrapper[4690]: I1203 13:34:21.386368 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l5z6x" event={"ID":"7d64a510-19ae-40a6-bf85-e538894ef917","Type":"ContainerStarted","Data":"42ce4aaa41870899489aca285ea55e26f94683fdb30313e528860668cbbcf385"} Dec 03 13:34:22 crc kubenswrapper[4690]: I1203 13:34:22.401242 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerStarted","Data":"e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450"} Dec 03 13:34:28 crc kubenswrapper[4690]: I1203 13:34:28.978068 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:34:29 crc kubenswrapper[4690]: I1203 13:34:29.092156 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-gm28d"] Dec 03 13:34:29 crc kubenswrapper[4690]: I1203 13:34:29.092949 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" podUID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" containerName="dnsmasq-dns" containerID="cri-o://6d0f45e48fb229e3a9ff169b79841cc3548277162151de824334818215f6cb23" gracePeriod=10 Dec 03 13:34:29 crc kubenswrapper[4690]: I1203 13:34:29.497745 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerStarted","Data":"0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662"} Dec 03 13:34:29 crc kubenswrapper[4690]: I1203 13:34:29.500466 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l5z6x" event={"ID":"7d64a510-19ae-40a6-bf85-e538894ef917","Type":"ContainerStarted","Data":"5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d"} Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.518730 4690 generic.go:334] "Generic (PLEG): container finished" podID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" containerID="6d0f45e48fb229e3a9ff169b79841cc3548277162151de824334818215f6cb23" exitCode=0 Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.518943 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" event={"ID":"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b","Type":"ContainerDied","Data":"6d0f45e48fb229e3a9ff169b79841cc3548277162151de824334818215f6cb23"} Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.564196 4690 generic.go:334] "Generic (PLEG): container finished" podID="7d64a510-19ae-40a6-bf85-e538894ef917" containerID="5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d" exitCode=0 Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.564245 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l5z6x" event={"ID":"7d64a510-19ae-40a6-bf85-e538894ef917","Type":"ContainerDied","Data":"5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d"} Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.805101 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.974325 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-sb\") pod \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.974495 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-svc\") pod \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.974557 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-config\") pod \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.974586 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-nb\") pod \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.974625 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-swift-storage-0\") pod \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.974683 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st47z\" (UniqueName: \"kubernetes.io/projected/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-kube-api-access-st47z\") pod \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\" (UID: \"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b\") " Dec 03 13:34:30 crc kubenswrapper[4690]: I1203 13:34:30.980666 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-kube-api-access-st47z" (OuterVolumeSpecName: "kube-api-access-st47z") pod "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" (UID: "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b"). InnerVolumeSpecName "kube-api-access-st47z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.029040 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" (UID: "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.030072 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" (UID: "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.033177 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" (UID: "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.034824 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-config" (OuterVolumeSpecName: "config") pod "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" (UID: "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.040625 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" (UID: "f4750fc4-e086-4be9-a5e8-aa7c0f778f6b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.078036 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.078077 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.078089 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.078102 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.078114 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st47z\" (UniqueName: \"kubernetes.io/projected/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-kube-api-access-st47z\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.078126 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.578079 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerStarted","Data":"012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181"} Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.578486 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.580887 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" event={"ID":"f4750fc4-e086-4be9-a5e8-aa7c0f778f6b","Type":"ContainerDied","Data":"84393b150ef0c8b61ef0aa3183f57559203ffceac415ea6c2313aece80bec53f"} Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.580949 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-gm28d" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.580991 4690 scope.go:117] "RemoveContainer" containerID="6d0f45e48fb229e3a9ff169b79841cc3548277162151de824334818215f6cb23" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.608891 4690 scope.go:117] "RemoveContainer" containerID="c7258d23bd6f397b55c6a1061f67a8ed53fe23e846e619f5e095335cd053ffd6" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.633388 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.792531498 podStartE2EDuration="13.633359274s" podCreationTimestamp="2025-12-03 13:34:18 +0000 UTC" firstStartedPulling="2025-12-03 13:34:19.383779978 +0000 UTC m=+1505.364700421" lastFinishedPulling="2025-12-03 13:34:30.224607764 +0000 UTC m=+1516.205528197" observedRunningTime="2025-12-03 13:34:31.622078047 +0000 UTC m=+1517.602998490" watchObservedRunningTime="2025-12-03 13:34:31.633359274 +0000 UTC m=+1517.614279707" Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.648712 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-gm28d"] Dec 03 13:34:31 crc kubenswrapper[4690]: I1203 13:34:31.659206 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-gm28d"] Dec 03 13:34:32 crc kubenswrapper[4690]: I1203 13:34:32.332397 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" path="/var/lib/kubelet/pods/f4750fc4-e086-4be9-a5e8-aa7c0f778f6b/volumes" Dec 03 13:34:32 crc kubenswrapper[4690]: I1203 13:34:32.595426 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l5z6x" event={"ID":"7d64a510-19ae-40a6-bf85-e538894ef917","Type":"ContainerStarted","Data":"5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc"} Dec 03 13:34:32 crc kubenswrapper[4690]: I1203 13:34:32.622906 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l5z6x" podStartSLOduration=3.6159746950000002 podStartE2EDuration="13.622862674s" podCreationTimestamp="2025-12-03 13:34:19 +0000 UTC" firstStartedPulling="2025-12-03 13:34:21.389375321 +0000 UTC m=+1507.370295754" lastFinishedPulling="2025-12-03 13:34:31.39626328 +0000 UTC m=+1517.377183733" observedRunningTime="2025-12-03 13:34:32.612716027 +0000 UTC m=+1518.593636460" watchObservedRunningTime="2025-12-03 13:34:32.622862674 +0000 UTC m=+1518.603783107" Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.890369 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-54g4r"] Dec 03 13:34:35 crc kubenswrapper[4690]: E1203 13:34:35.891590 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" containerName="dnsmasq-dns" Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.891609 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" containerName="dnsmasq-dns" Dec 03 13:34:35 crc kubenswrapper[4690]: E1203 13:34:35.891621 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" containerName="init" Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.891629 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" containerName="init" Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.891833 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4750fc4-e086-4be9-a5e8-aa7c0f778f6b" containerName="dnsmasq-dns" Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.893562 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.938295 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-54g4r"] Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.997086 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-utilities\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.997172 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-catalog-content\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:35 crc kubenswrapper[4690]: I1203 13:34:35.997327 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdbhj\" (UniqueName: \"kubernetes.io/projected/a118e628-1d51-4152-9898-f1c53eb2f471-kube-api-access-wdbhj\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.099430 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdbhj\" (UniqueName: \"kubernetes.io/projected/a118e628-1d51-4152-9898-f1c53eb2f471-kube-api-access-wdbhj\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.099545 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-utilities\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.099593 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-catalog-content\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.100235 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-utilities\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.100258 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-catalog-content\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.123495 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdbhj\" (UniqueName: \"kubernetes.io/projected/a118e628-1d51-4152-9898-f1c53eb2f471-kube-api-access-wdbhj\") pod \"redhat-operators-54g4r\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.229683 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.586458 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.589204 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="ceilometer-central-agent" containerID="cri-o://7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da" gracePeriod=30 Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.589329 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="proxy-httpd" containerID="cri-o://012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181" gracePeriod=30 Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.589338 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="sg-core" containerID="cri-o://0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662" gracePeriod=30 Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.589383 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="ceilometer-notification-agent" containerID="cri-o://e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450" gracePeriod=30 Dec 03 13:34:36 crc kubenswrapper[4690]: I1203 13:34:36.755155 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-54g4r"] Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.657849 4690 generic.go:334] "Generic (PLEG): container finished" podID="01b61175-58d3-4291-9b9f-c791281fe495" containerID="012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181" exitCode=0 Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.658276 4690 generic.go:334] "Generic (PLEG): container finished" podID="01b61175-58d3-4291-9b9f-c791281fe495" containerID="0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662" exitCode=2 Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.658288 4690 generic.go:334] "Generic (PLEG): container finished" podID="01b61175-58d3-4291-9b9f-c791281fe495" containerID="7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da" exitCode=0 Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.657925 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerDied","Data":"012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181"} Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.658374 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerDied","Data":"0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662"} Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.658390 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerDied","Data":"7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da"} Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.662392 4690 generic.go:334] "Generic (PLEG): container finished" podID="a118e628-1d51-4152-9898-f1c53eb2f471" containerID="dade26139e1996104cfdc8039cb67c7271c2efc3dc3997f266272aafb680e35e" exitCode=0 Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.662560 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54g4r" event={"ID":"a118e628-1d51-4152-9898-f1c53eb2f471","Type":"ContainerDied","Data":"dade26139e1996104cfdc8039cb67c7271c2efc3dc3997f266272aafb680e35e"} Dec 03 13:34:37 crc kubenswrapper[4690]: I1203 13:34:37.662667 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54g4r" event={"ID":"a118e628-1d51-4152-9898-f1c53eb2f471","Type":"ContainerStarted","Data":"5db7d7a3cd28941eb12084bdb4544bce4ec8a604c2116c5e49cd34a56853a657"} Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.233314 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.378125 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-config-data\") pod \"01b61175-58d3-4291-9b9f-c791281fe495\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.378460 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-scripts\") pod \"01b61175-58d3-4291-9b9f-c791281fe495\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.378508 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-combined-ca-bundle\") pod \"01b61175-58d3-4291-9b9f-c791281fe495\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.378537 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-sg-core-conf-yaml\") pod \"01b61175-58d3-4291-9b9f-c791281fe495\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.378564 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-run-httpd\") pod \"01b61175-58d3-4291-9b9f-c791281fe495\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.378625 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-log-httpd\") pod \"01b61175-58d3-4291-9b9f-c791281fe495\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.378702 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llcrf\" (UniqueName: \"kubernetes.io/projected/01b61175-58d3-4291-9b9f-c791281fe495-kube-api-access-llcrf\") pod \"01b61175-58d3-4291-9b9f-c791281fe495\" (UID: \"01b61175-58d3-4291-9b9f-c791281fe495\") " Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.380716 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "01b61175-58d3-4291-9b9f-c791281fe495" (UID: "01b61175-58d3-4291-9b9f-c791281fe495"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.381479 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "01b61175-58d3-4291-9b9f-c791281fe495" (UID: "01b61175-58d3-4291-9b9f-c791281fe495"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.401958 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-scripts" (OuterVolumeSpecName: "scripts") pod "01b61175-58d3-4291-9b9f-c791281fe495" (UID: "01b61175-58d3-4291-9b9f-c791281fe495"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.413878 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01b61175-58d3-4291-9b9f-c791281fe495-kube-api-access-llcrf" (OuterVolumeSpecName: "kube-api-access-llcrf") pod "01b61175-58d3-4291-9b9f-c791281fe495" (UID: "01b61175-58d3-4291-9b9f-c791281fe495"). InnerVolumeSpecName "kube-api-access-llcrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.425516 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "01b61175-58d3-4291-9b9f-c791281fe495" (UID: "01b61175-58d3-4291-9b9f-c791281fe495"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.481365 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.481702 4690 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.481792 4690 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.481924 4690 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01b61175-58d3-4291-9b9f-c791281fe495-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.481967 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llcrf\" (UniqueName: \"kubernetes.io/projected/01b61175-58d3-4291-9b9f-c791281fe495-kube-api-access-llcrf\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.504649 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01b61175-58d3-4291-9b9f-c791281fe495" (UID: "01b61175-58d3-4291-9b9f-c791281fe495"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.505164 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-config-data" (OuterVolumeSpecName: "config-data") pod "01b61175-58d3-4291-9b9f-c791281fe495" (UID: "01b61175-58d3-4291-9b9f-c791281fe495"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.584903 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.584958 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01b61175-58d3-4291-9b9f-c791281fe495-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.687242 4690 generic.go:334] "Generic (PLEG): container finished" podID="a118e628-1d51-4152-9898-f1c53eb2f471" containerID="341002d7e13b4af8c0874b0ff1f556310ae2b490d1209f0f004d79588f206433" exitCode=0 Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.687403 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54g4r" event={"ID":"a118e628-1d51-4152-9898-f1c53eb2f471","Type":"ContainerDied","Data":"341002d7e13b4af8c0874b0ff1f556310ae2b490d1209f0f004d79588f206433"} Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.697250 4690 generic.go:334] "Generic (PLEG): container finished" podID="01b61175-58d3-4291-9b9f-c791281fe495" containerID="e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450" exitCode=0 Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.697354 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerDied","Data":"e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450"} Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.698081 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01b61175-58d3-4291-9b9f-c791281fe495","Type":"ContainerDied","Data":"0120fa55e3fdb606c0b15ec769e06b0aa7854fd476de2513e57e2e74bfec82ae"} Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.698162 4690 scope.go:117] "RemoveContainer" containerID="012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.697374 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.740218 4690 scope.go:117] "RemoveContainer" containerID="0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.753116 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.786603 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.793621 4690 scope.go:117] "RemoveContainer" containerID="e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.823853 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:39 crc kubenswrapper[4690]: E1203 13:34:39.825126 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="sg-core" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.825157 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="sg-core" Dec 03 13:34:39 crc kubenswrapper[4690]: E1203 13:34:39.825218 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="ceilometer-central-agent" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.825227 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="ceilometer-central-agent" Dec 03 13:34:39 crc kubenswrapper[4690]: E1203 13:34:39.825246 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="proxy-httpd" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.825257 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="proxy-httpd" Dec 03 13:34:39 crc kubenswrapper[4690]: E1203 13:34:39.825285 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="ceilometer-notification-agent" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.825292 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="ceilometer-notification-agent" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.825820 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="ceilometer-central-agent" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.825856 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="sg-core" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.825908 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="ceilometer-notification-agent" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.825933 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="01b61175-58d3-4291-9b9f-c791281fe495" containerName="proxy-httpd" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.834700 4690 scope.go:117] "RemoveContainer" containerID="7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.836193 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.839258 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.839367 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.847503 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.870844 4690 scope.go:117] "RemoveContainer" containerID="012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181" Dec 03 13:34:39 crc kubenswrapper[4690]: E1203 13:34:39.871489 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181\": container with ID starting with 012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181 not found: ID does not exist" containerID="012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.871535 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181"} err="failed to get container status \"012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181\": rpc error: code = NotFound desc = could not find container \"012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181\": container with ID starting with 012afa3366aeeedc99de315716243d850cb676d412009232852b8751e0ba9181 not found: ID does not exist" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.871565 4690 scope.go:117] "RemoveContainer" containerID="0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662" Dec 03 13:34:39 crc kubenswrapper[4690]: E1203 13:34:39.872105 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662\": container with ID starting with 0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662 not found: ID does not exist" containerID="0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.872169 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662"} err="failed to get container status \"0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662\": rpc error: code = NotFound desc = could not find container \"0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662\": container with ID starting with 0db16d4789e497d48ec78084882297a9047c4f9142af60409229b894117d4662 not found: ID does not exist" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.872209 4690 scope.go:117] "RemoveContainer" containerID="e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450" Dec 03 13:34:39 crc kubenswrapper[4690]: E1203 13:34:39.872737 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450\": container with ID starting with e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450 not found: ID does not exist" containerID="e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.872766 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450"} err="failed to get container status \"e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450\": rpc error: code = NotFound desc = could not find container \"e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450\": container with ID starting with e626ccbf1826c9426da9ac76f3bb365427bb927d0e37fdedd0e6e5820afff450 not found: ID does not exist" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.872782 4690 scope.go:117] "RemoveContainer" containerID="7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da" Dec 03 13:34:39 crc kubenswrapper[4690]: E1203 13:34:39.873136 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da\": container with ID starting with 7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da not found: ID does not exist" containerID="7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da" Dec 03 13:34:39 crc kubenswrapper[4690]: I1203 13:34:39.873163 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da"} err="failed to get container status \"7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da\": rpc error: code = NotFound desc = could not find container \"7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da\": container with ID starting with 7c8360ad07bc5fe31d266762d83e28035309a00fe675c75982f8d4c85fdb46da not found: ID does not exist" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.009539 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.009601 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-log-httpd\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.009683 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-config-data\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.009734 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.009761 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x2dr\" (UniqueName: \"kubernetes.io/projected/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-kube-api-access-2x2dr\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.009778 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-scripts\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.009801 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-run-httpd\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.090762 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.090813 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.111884 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-log-httpd\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.112355 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-log-httpd\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.112406 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-config-data\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.112504 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.112533 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x2dr\" (UniqueName: \"kubernetes.io/projected/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-kube-api-access-2x2dr\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.112565 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-scripts\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.112596 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-run-httpd\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.112749 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.113212 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-run-httpd\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.116734 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-scripts\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.117175 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-config-data\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.118319 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.124562 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.131383 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x2dr\" (UniqueName: \"kubernetes.io/projected/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-kube-api-access-2x2dr\") pod \"ceilometer-0\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.149689 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.163584 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.328099 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01b61175-58d3-4291-9b9f-c791281fe495" path="/var/lib/kubelet/pods/01b61175-58d3-4291-9b9f-c791281fe495/volumes" Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.644475 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.714860 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerStarted","Data":"31ba55e36ad545dabc4affac436707fc0df119042fa4850aa6857797cea6d444"} Dec 03 13:34:40 crc kubenswrapper[4690]: I1203 13:34:40.801743 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:41 crc kubenswrapper[4690]: I1203 13:34:41.264528 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l5z6x"] Dec 03 13:34:42 crc kubenswrapper[4690]: I1203 13:34:42.740905 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54g4r" event={"ID":"a118e628-1d51-4152-9898-f1c53eb2f471","Type":"ContainerStarted","Data":"b489ec7a091b54bb64d5d7811da5c87dbac3fc91d3084d58ea3edcb81b336fa7"} Dec 03 13:34:42 crc kubenswrapper[4690]: I1203 13:34:42.744792 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l5z6x" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" containerName="registry-server" containerID="cri-o://5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc" gracePeriod=2 Dec 03 13:34:42 crc kubenswrapper[4690]: I1203 13:34:42.744943 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerStarted","Data":"44255d27ab5f10c96b4a3fc74811b968b3b155c4c7baa87b82878e14b3c08cd3"} Dec 03 13:34:42 crc kubenswrapper[4690]: I1203 13:34:42.784798 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-54g4r" podStartSLOduration=4.910377802 podStartE2EDuration="7.784780172s" podCreationTimestamp="2025-12-03 13:34:35 +0000 UTC" firstStartedPulling="2025-12-03 13:34:37.664390768 +0000 UTC m=+1523.645311211" lastFinishedPulling="2025-12-03 13:34:40.538793148 +0000 UTC m=+1526.519713581" observedRunningTime="2025-12-03 13:34:42.783289422 +0000 UTC m=+1528.764209855" watchObservedRunningTime="2025-12-03 13:34:42.784780172 +0000 UTC m=+1528.765700605" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.367681 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.491540 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-utilities\") pod \"7d64a510-19ae-40a6-bf85-e538894ef917\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.491783 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-catalog-content\") pod \"7d64a510-19ae-40a6-bf85-e538894ef917\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.492088 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blsxv\" (UniqueName: \"kubernetes.io/projected/7d64a510-19ae-40a6-bf85-e538894ef917-kube-api-access-blsxv\") pod \"7d64a510-19ae-40a6-bf85-e538894ef917\" (UID: \"7d64a510-19ae-40a6-bf85-e538894ef917\") " Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.492485 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-utilities" (OuterVolumeSpecName: "utilities") pod "7d64a510-19ae-40a6-bf85-e538894ef917" (UID: "7d64a510-19ae-40a6-bf85-e538894ef917"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.492740 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.503086 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d64a510-19ae-40a6-bf85-e538894ef917-kube-api-access-blsxv" (OuterVolumeSpecName: "kube-api-access-blsxv") pod "7d64a510-19ae-40a6-bf85-e538894ef917" (UID: "7d64a510-19ae-40a6-bf85-e538894ef917"). InnerVolumeSpecName "kube-api-access-blsxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.548276 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d64a510-19ae-40a6-bf85-e538894ef917" (UID: "7d64a510-19ae-40a6-bf85-e538894ef917"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.594771 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d64a510-19ae-40a6-bf85-e538894ef917-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.594817 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blsxv\" (UniqueName: \"kubernetes.io/projected/7d64a510-19ae-40a6-bf85-e538894ef917-kube-api-access-blsxv\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.765322 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.771512 4690 generic.go:334] "Generic (PLEG): container finished" podID="7d64a510-19ae-40a6-bf85-e538894ef917" containerID="5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc" exitCode=0 Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.772856 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l5z6x" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.775294 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l5z6x" event={"ID":"7d64a510-19ae-40a6-bf85-e538894ef917","Type":"ContainerDied","Data":"5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc"} Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.775396 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l5z6x" event={"ID":"7d64a510-19ae-40a6-bf85-e538894ef917","Type":"ContainerDied","Data":"42ce4aaa41870899489aca285ea55e26f94683fdb30313e528860668cbbcf385"} Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.775426 4690 scope.go:117] "RemoveContainer" containerID="5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.862664 4690 scope.go:117] "RemoveContainer" containerID="5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d" Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.904675 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l5z6x"] Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.928769 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l5z6x"] Dec 03 13:34:43 crc kubenswrapper[4690]: I1203 13:34:43.952297 4690 scope.go:117] "RemoveContainer" containerID="40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577" Dec 03 13:34:44 crc kubenswrapper[4690]: I1203 13:34:44.021700 4690 scope.go:117] "RemoveContainer" containerID="5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc" Dec 03 13:34:44 crc kubenswrapper[4690]: E1203 13:34:44.022935 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc\": container with ID starting with 5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc not found: ID does not exist" containerID="5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc" Dec 03 13:34:44 crc kubenswrapper[4690]: I1203 13:34:44.022972 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc"} err="failed to get container status \"5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc\": rpc error: code = NotFound desc = could not find container \"5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc\": container with ID starting with 5e6c9f17cddcd8ddba728433018ad9b9d468928ef9b395ee62ab3fd837bdf0fc not found: ID does not exist" Dec 03 13:34:44 crc kubenswrapper[4690]: I1203 13:34:44.022996 4690 scope.go:117] "RemoveContainer" containerID="5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d" Dec 03 13:34:44 crc kubenswrapper[4690]: E1203 13:34:44.024057 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d\": container with ID starting with 5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d not found: ID does not exist" containerID="5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d" Dec 03 13:34:44 crc kubenswrapper[4690]: I1203 13:34:44.024081 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d"} err="failed to get container status \"5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d\": rpc error: code = NotFound desc = could not find container \"5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d\": container with ID starting with 5da85496d71ad4689a5adbbff3cc3f929842e1398c5e67b0e4031e961504be8d not found: ID does not exist" Dec 03 13:34:44 crc kubenswrapper[4690]: I1203 13:34:44.024097 4690 scope.go:117] "RemoveContainer" containerID="40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577" Dec 03 13:34:44 crc kubenswrapper[4690]: E1203 13:34:44.024857 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577\": container with ID starting with 40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577 not found: ID does not exist" containerID="40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577" Dec 03 13:34:44 crc kubenswrapper[4690]: I1203 13:34:44.024895 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577"} err="failed to get container status \"40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577\": rpc error: code = NotFound desc = could not find container \"40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577\": container with ID starting with 40f488e2e9d6854bbf44f015991942f385e7315ad5f91bacfe55ba751e738577 not found: ID does not exist" Dec 03 13:34:44 crc kubenswrapper[4690]: I1203 13:34:44.326528 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" path="/var/lib/kubelet/pods/7d64a510-19ae-40a6-bf85-e538894ef917/volumes" Dec 03 13:34:44 crc kubenswrapper[4690]: I1203 13:34:44.786004 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerStarted","Data":"45cd8e316e6636de21df426cb31d01e567b19ad8140be8156f77dfd45ea4b35c"} Dec 03 13:34:46 crc kubenswrapper[4690]: I1203 13:34:46.229907 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:46 crc kubenswrapper[4690]: I1203 13:34:46.231727 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:46 crc kubenswrapper[4690]: I1203 13:34:46.372162 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:34:47 crc kubenswrapper[4690]: I1203 13:34:47.056576 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d87869f9-cmzfc" Dec 03 13:34:47 crc kubenswrapper[4690]: I1203 13:34:47.134746 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65497df674-vnvht"] Dec 03 13:34:47 crc kubenswrapper[4690]: I1203 13:34:47.135028 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65497df674-vnvht" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerName="neutron-api" containerID="cri-o://be72616885f72bda5f6b02089fdc438a90dee30aa3513d38e8768270f11c787f" gracePeriod=30 Dec 03 13:34:47 crc kubenswrapper[4690]: I1203 13:34:47.135497 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65497df674-vnvht" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerName="neutron-httpd" containerID="cri-o://a5717985d3761a304110d932fdcd5d6ba72e5ce8b97174b2086f03ef0883d685" gracePeriod=30 Dec 03 13:34:47 crc kubenswrapper[4690]: I1203 13:34:47.295165 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-54g4r" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="registry-server" probeResult="failure" output=< Dec 03 13:34:47 crc kubenswrapper[4690]: timeout: failed to connect service ":50051" within 1s Dec 03 13:34:47 crc kubenswrapper[4690]: > Dec 03 13:34:48 crc kubenswrapper[4690]: I1203 13:34:48.847581 4690 generic.go:334] "Generic (PLEG): container finished" podID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerID="a5717985d3761a304110d932fdcd5d6ba72e5ce8b97174b2086f03ef0883d685" exitCode=0 Dec 03 13:34:48 crc kubenswrapper[4690]: I1203 13:34:48.847607 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65497df674-vnvht" event={"ID":"5ce567ee-468c-4f48-835d-a09de85f5d93","Type":"ContainerDied","Data":"a5717985d3761a304110d932fdcd5d6ba72e5ce8b97174b2086f03ef0883d685"} Dec 03 13:34:52 crc kubenswrapper[4690]: I1203 13:34:52.896305 4690 generic.go:334] "Generic (PLEG): container finished" podID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerID="be72616885f72bda5f6b02089fdc438a90dee30aa3513d38e8768270f11c787f" exitCode=0 Dec 03 13:34:52 crc kubenswrapper[4690]: I1203 13:34:52.897129 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65497df674-vnvht" event={"ID":"5ce567ee-468c-4f48-835d-a09de85f5d93","Type":"ContainerDied","Data":"be72616885f72bda5f6b02089fdc438a90dee30aa3513d38e8768270f11c787f"} Dec 03 13:34:56 crc kubenswrapper[4690]: I1203 13:34:56.326198 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:56 crc kubenswrapper[4690]: I1203 13:34:56.376241 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:34:56 crc kubenswrapper[4690]: I1203 13:34:56.563527 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-54g4r"] Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.123790 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.226988 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-config\") pod \"5ce567ee-468c-4f48-835d-a09de85f5d93\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.227092 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgnn9\" (UniqueName: \"kubernetes.io/projected/5ce567ee-468c-4f48-835d-a09de85f5d93-kube-api-access-vgnn9\") pod \"5ce567ee-468c-4f48-835d-a09de85f5d93\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.227228 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-combined-ca-bundle\") pod \"5ce567ee-468c-4f48-835d-a09de85f5d93\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.227294 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-ovndb-tls-certs\") pod \"5ce567ee-468c-4f48-835d-a09de85f5d93\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.227318 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-httpd-config\") pod \"5ce567ee-468c-4f48-835d-a09de85f5d93\" (UID: \"5ce567ee-468c-4f48-835d-a09de85f5d93\") " Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.236556 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ce567ee-468c-4f48-835d-a09de85f5d93-kube-api-access-vgnn9" (OuterVolumeSpecName: "kube-api-access-vgnn9") pod "5ce567ee-468c-4f48-835d-a09de85f5d93" (UID: "5ce567ee-468c-4f48-835d-a09de85f5d93"). InnerVolumeSpecName "kube-api-access-vgnn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.256829 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5ce567ee-468c-4f48-835d-a09de85f5d93" (UID: "5ce567ee-468c-4f48-835d-a09de85f5d93"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.287509 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-config" (OuterVolumeSpecName: "config") pod "5ce567ee-468c-4f48-835d-a09de85f5d93" (UID: "5ce567ee-468c-4f48-835d-a09de85f5d93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.304031 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ce567ee-468c-4f48-835d-a09de85f5d93" (UID: "5ce567ee-468c-4f48-835d-a09de85f5d93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.332957 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.332993 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgnn9\" (UniqueName: \"kubernetes.io/projected/5ce567ee-468c-4f48-835d-a09de85f5d93-kube-api-access-vgnn9\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.333006 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.333016 4690 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.360914 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5ce567ee-468c-4f48-835d-a09de85f5d93" (UID: "5ce567ee-468c-4f48-835d-a09de85f5d93"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:34:57 crc kubenswrapper[4690]: I1203 13:34:57.435217 4690 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ce567ee-468c-4f48-835d-a09de85f5d93-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:34:58 crc kubenswrapper[4690]: I1203 13:34:58.098412 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-54g4r" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="registry-server" containerID="cri-o://b489ec7a091b54bb64d5d7811da5c87dbac3fc91d3084d58ea3edcb81b336fa7" gracePeriod=2 Dec 03 13:34:58 crc kubenswrapper[4690]: I1203 13:34:58.099354 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65497df674-vnvht" Dec 03 13:34:58 crc kubenswrapper[4690]: I1203 13:34:58.099410 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65497df674-vnvht" event={"ID":"5ce567ee-468c-4f48-835d-a09de85f5d93","Type":"ContainerDied","Data":"d97a168ca3980ef383c6447c5f7c511c70cd172a3752a6d3089e3d5599bf82f5"} Dec 03 13:34:58 crc kubenswrapper[4690]: I1203 13:34:58.100430 4690 scope.go:117] "RemoveContainer" containerID="a5717985d3761a304110d932fdcd5d6ba72e5ce8b97174b2086f03ef0883d685" Dec 03 13:34:58 crc kubenswrapper[4690]: I1203 13:34:58.146122 4690 scope.go:117] "RemoveContainer" containerID="be72616885f72bda5f6b02089fdc438a90dee30aa3513d38e8768270f11c787f" Dec 03 13:34:58 crc kubenswrapper[4690]: I1203 13:34:58.155270 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65497df674-vnvht"] Dec 03 13:34:58 crc kubenswrapper[4690]: I1203 13:34:58.166614 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-65497df674-vnvht"] Dec 03 13:34:58 crc kubenswrapper[4690]: I1203 13:34:58.333267 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" path="/var/lib/kubelet/pods/5ce567ee-468c-4f48-835d-a09de85f5d93/volumes" Dec 03 13:34:59 crc kubenswrapper[4690]: I1203 13:34:59.114309 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerStarted","Data":"95172ea3e37dbdc4f54830ad04cd06abe1f20fd9e0eb60fd78fb9765d7a8a47a"} Dec 03 13:35:01 crc kubenswrapper[4690]: I1203 13:35:01.136276 4690 generic.go:334] "Generic (PLEG): container finished" podID="a118e628-1d51-4152-9898-f1c53eb2f471" containerID="b489ec7a091b54bb64d5d7811da5c87dbac3fc91d3084d58ea3edcb81b336fa7" exitCode=0 Dec 03 13:35:01 crc kubenswrapper[4690]: I1203 13:35:01.136356 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54g4r" event={"ID":"a118e628-1d51-4152-9898-f1c53eb2f471","Type":"ContainerDied","Data":"b489ec7a091b54bb64d5d7811da5c87dbac3fc91d3084d58ea3edcb81b336fa7"} Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.291739 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.351859 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-utilities\") pod \"a118e628-1d51-4152-9898-f1c53eb2f471\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.351997 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-catalog-content\") pod \"a118e628-1d51-4152-9898-f1c53eb2f471\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.352239 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdbhj\" (UniqueName: \"kubernetes.io/projected/a118e628-1d51-4152-9898-f1c53eb2f471-kube-api-access-wdbhj\") pod \"a118e628-1d51-4152-9898-f1c53eb2f471\" (UID: \"a118e628-1d51-4152-9898-f1c53eb2f471\") " Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.354302 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-utilities" (OuterVolumeSpecName: "utilities") pod "a118e628-1d51-4152-9898-f1c53eb2f471" (UID: "a118e628-1d51-4152-9898-f1c53eb2f471"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.363964 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a118e628-1d51-4152-9898-f1c53eb2f471-kube-api-access-wdbhj" (OuterVolumeSpecName: "kube-api-access-wdbhj") pod "a118e628-1d51-4152-9898-f1c53eb2f471" (UID: "a118e628-1d51-4152-9898-f1c53eb2f471"). InnerVolumeSpecName "kube-api-access-wdbhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.454886 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdbhj\" (UniqueName: \"kubernetes.io/projected/a118e628-1d51-4152-9898-f1c53eb2f471-kube-api-access-wdbhj\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.454926 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.521333 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a118e628-1d51-4152-9898-f1c53eb2f471" (UID: "a118e628-1d51-4152-9898-f1c53eb2f471"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:02 crc kubenswrapper[4690]: I1203 13:35:02.558659 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a118e628-1d51-4152-9898-f1c53eb2f471-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:03 crc kubenswrapper[4690]: I1203 13:35:03.171166 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54g4r" event={"ID":"a118e628-1d51-4152-9898-f1c53eb2f471","Type":"ContainerDied","Data":"5db7d7a3cd28941eb12084bdb4544bce4ec8a604c2116c5e49cd34a56853a657"} Dec 03 13:35:03 crc kubenswrapper[4690]: I1203 13:35:03.171219 4690 scope.go:117] "RemoveContainer" containerID="b489ec7a091b54bb64d5d7811da5c87dbac3fc91d3084d58ea3edcb81b336fa7" Dec 03 13:35:03 crc kubenswrapper[4690]: I1203 13:35:03.171339 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54g4r" Dec 03 13:35:03 crc kubenswrapper[4690]: I1203 13:35:03.210315 4690 scope.go:117] "RemoveContainer" containerID="341002d7e13b4af8c0874b0ff1f556310ae2b490d1209f0f004d79588f206433" Dec 03 13:35:03 crc kubenswrapper[4690]: I1203 13:35:03.217875 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-54g4r"] Dec 03 13:35:03 crc kubenswrapper[4690]: I1203 13:35:03.232928 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-54g4r"] Dec 03 13:35:03 crc kubenswrapper[4690]: I1203 13:35:03.247689 4690 scope.go:117] "RemoveContainer" containerID="dade26139e1996104cfdc8039cb67c7271c2efc3dc3997f266272aafb680e35e" Dec 03 13:35:04 crc kubenswrapper[4690]: I1203 13:35:04.342025 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" path="/var/lib/kubelet/pods/a118e628-1d51-4152-9898-f1c53eb2f471/volumes" Dec 03 13:35:08 crc kubenswrapper[4690]: I1203 13:35:08.229514 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerStarted","Data":"e5ebb4907d1025c203c301da08555282a651966549e0b7275e8be1784bdf9d89"} Dec 03 13:35:08 crc kubenswrapper[4690]: I1203 13:35:08.230234 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:35:08 crc kubenswrapper[4690]: I1203 13:35:08.229848 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="proxy-httpd" containerID="cri-o://e5ebb4907d1025c203c301da08555282a651966549e0b7275e8be1784bdf9d89" gracePeriod=30 Dec 03 13:35:08 crc kubenswrapper[4690]: I1203 13:35:08.229769 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="ceilometer-central-agent" containerID="cri-o://44255d27ab5f10c96b4a3fc74811b968b3b155c4c7baa87b82878e14b3c08cd3" gracePeriod=30 Dec 03 13:35:08 crc kubenswrapper[4690]: I1203 13:35:08.229899 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="sg-core" containerID="cri-o://95172ea3e37dbdc4f54830ad04cd06abe1f20fd9e0eb60fd78fb9765d7a8a47a" gracePeriod=30 Dec 03 13:35:08 crc kubenswrapper[4690]: I1203 13:35:08.229899 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="ceilometer-notification-agent" containerID="cri-o://45cd8e316e6636de21df426cb31d01e567b19ad8140be8156f77dfd45ea4b35c" gracePeriod=30 Dec 03 13:35:09 crc kubenswrapper[4690]: I1203 13:35:09.277321 4690 generic.go:334] "Generic (PLEG): container finished" podID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerID="e5ebb4907d1025c203c301da08555282a651966549e0b7275e8be1784bdf9d89" exitCode=0 Dec 03 13:35:09 crc kubenswrapper[4690]: I1203 13:35:09.277747 4690 generic.go:334] "Generic (PLEG): container finished" podID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerID="95172ea3e37dbdc4f54830ad04cd06abe1f20fd9e0eb60fd78fb9765d7a8a47a" exitCode=2 Dec 03 13:35:09 crc kubenswrapper[4690]: I1203 13:35:09.277760 4690 generic.go:334] "Generic (PLEG): container finished" podID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerID="45cd8e316e6636de21df426cb31d01e567b19ad8140be8156f77dfd45ea4b35c" exitCode=0 Dec 03 13:35:09 crc kubenswrapper[4690]: I1203 13:35:09.277537 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerDied","Data":"e5ebb4907d1025c203c301da08555282a651966549e0b7275e8be1784bdf9d89"} Dec 03 13:35:09 crc kubenswrapper[4690]: I1203 13:35:09.277799 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerDied","Data":"95172ea3e37dbdc4f54830ad04cd06abe1f20fd9e0eb60fd78fb9765d7a8a47a"} Dec 03 13:35:09 crc kubenswrapper[4690]: I1203 13:35:09.277813 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerDied","Data":"45cd8e316e6636de21df426cb31d01e567b19ad8140be8156f77dfd45ea4b35c"} Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.290585 4690 generic.go:334] "Generic (PLEG): container finished" podID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerID="44255d27ab5f10c96b4a3fc74811b968b3b155c4c7baa87b82878e14b3c08cd3" exitCode=0 Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.290782 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerDied","Data":"44255d27ab5f10c96b4a3fc74811b968b3b155c4c7baa87b82878e14b3c08cd3"} Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.291074 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3f4cdc2-0994-4a59-9a7a-c6680f10d359","Type":"ContainerDied","Data":"31ba55e36ad545dabc4affac436707fc0df119042fa4850aa6857797cea6d444"} Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.291094 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31ba55e36ad545dabc4affac436707fc0df119042fa4850aa6857797cea6d444" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.293200 4690 generic.go:334] "Generic (PLEG): container finished" podID="43129653-d1cd-4ced-b120-33be343c3d9d" containerID="41fa04cdec6857177bb3fc18b10b5e4e11a9beade25ddd077b1f275dc2a49aa8" exitCode=0 Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.293231 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" event={"ID":"43129653-d1cd-4ced-b120-33be343c3d9d","Type":"ContainerDied","Data":"41fa04cdec6857177bb3fc18b10b5e4e11a9beade25ddd077b1f275dc2a49aa8"} Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.320123 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.240181542 podStartE2EDuration="31.320101291s" podCreationTimestamp="2025-12-03 13:34:39 +0000 UTC" firstStartedPulling="2025-12-03 13:34:40.651340407 +0000 UTC m=+1526.632260840" lastFinishedPulling="2025-12-03 13:35:06.731260166 +0000 UTC m=+1552.712180589" observedRunningTime="2025-12-03 13:35:08.266243254 +0000 UTC m=+1554.247163697" watchObservedRunningTime="2025-12-03 13:35:10.320101291 +0000 UTC m=+1556.301021744" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.323540 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.327157 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-scripts\") pod \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.327357 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-log-httpd\") pod \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.327391 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-combined-ca-bundle\") pod \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.327438 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-config-data\") pod \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.327553 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x2dr\" (UniqueName: \"kubernetes.io/projected/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-kube-api-access-2x2dr\") pod \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.327586 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-run-httpd\") pod \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.327632 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-sg-core-conf-yaml\") pod \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\" (UID: \"d3f4cdc2-0994-4a59-9a7a-c6680f10d359\") " Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.328466 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d3f4cdc2-0994-4a59-9a7a-c6680f10d359" (UID: "d3f4cdc2-0994-4a59-9a7a-c6680f10d359"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.329164 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d3f4cdc2-0994-4a59-9a7a-c6680f10d359" (UID: "d3f4cdc2-0994-4a59-9a7a-c6680f10d359"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.335772 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-kube-api-access-2x2dr" (OuterVolumeSpecName: "kube-api-access-2x2dr") pod "d3f4cdc2-0994-4a59-9a7a-c6680f10d359" (UID: "d3f4cdc2-0994-4a59-9a7a-c6680f10d359"). InnerVolumeSpecName "kube-api-access-2x2dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.337515 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-scripts" (OuterVolumeSpecName: "scripts") pod "d3f4cdc2-0994-4a59-9a7a-c6680f10d359" (UID: "d3f4cdc2-0994-4a59-9a7a-c6680f10d359"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.365593 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d3f4cdc2-0994-4a59-9a7a-c6680f10d359" (UID: "d3f4cdc2-0994-4a59-9a7a-c6680f10d359"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.430823 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.431191 4690 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.431203 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x2dr\" (UniqueName: \"kubernetes.io/projected/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-kube-api-access-2x2dr\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.431215 4690 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.431224 4690 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.432102 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3f4cdc2-0994-4a59-9a7a-c6680f10d359" (UID: "d3f4cdc2-0994-4a59-9a7a-c6680f10d359"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.472227 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-config-data" (OuterVolumeSpecName: "config-data") pod "d3f4cdc2-0994-4a59-9a7a-c6680f10d359" (UID: "d3f4cdc2-0994-4a59-9a7a-c6680f10d359"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.535257 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:10 crc kubenswrapper[4690]: I1203 13:35:10.535301 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f4cdc2-0994-4a59-9a7a-c6680f10d359-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.302068 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.346510 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.357396 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.388815 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389366 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="registry-server" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389393 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="registry-server" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389409 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" containerName="extract-content" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389417 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" containerName="extract-content" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389433 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="proxy-httpd" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389443 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="proxy-httpd" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389469 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerName="neutron-api" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389477 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerName="neutron-api" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389495 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerName="neutron-httpd" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389503 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerName="neutron-httpd" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389521 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="ceilometer-central-agent" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389528 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="ceilometer-central-agent" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389541 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="ceilometer-notification-agent" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389550 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="ceilometer-notification-agent" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389569 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" containerName="registry-server" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389577 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" containerName="registry-server" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389589 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="extract-utilities" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389596 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="extract-utilities" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389611 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="sg-core" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389618 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="sg-core" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389627 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" containerName="extract-utilities" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389634 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" containerName="extract-utilities" Dec 03 13:35:11 crc kubenswrapper[4690]: E1203 13:35:11.389646 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="extract-content" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389653 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="extract-content" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389895 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="proxy-httpd" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389909 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a118e628-1d51-4152-9898-f1c53eb2f471" containerName="registry-server" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389924 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerName="neutron-httpd" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389944 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="ceilometer-central-agent" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389962 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce567ee-468c-4f48-835d-a09de85f5d93" containerName="neutron-api" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389975 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d64a510-19ae-40a6-bf85-e538894ef917" containerName="registry-server" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389987 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="ceilometer-notification-agent" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.389999 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" containerName="sg-core" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.394350 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.396818 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.398961 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.409184 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.456536 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.456649 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29lrc\" (UniqueName: \"kubernetes.io/projected/7fac3e34-0060-4311-ad94-762e417c578b-kube-api-access-29lrc\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.456745 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-log-httpd\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.456786 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.456816 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-scripts\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.456898 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-config-data\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.456937 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-run-httpd\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.559245 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.559611 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29lrc\" (UniqueName: \"kubernetes.io/projected/7fac3e34-0060-4311-ad94-762e417c578b-kube-api-access-29lrc\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.559692 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-log-httpd\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.559725 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.559748 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-scripts\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.559806 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-config-data\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.559838 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-run-httpd\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.560364 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-run-httpd\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.566742 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-log-httpd\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.569953 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.572483 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-config-data\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.577565 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-scripts\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.586614 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.589151 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29lrc\" (UniqueName: \"kubernetes.io/projected/7fac3e34-0060-4311-ad94-762e417c578b-kube-api-access-29lrc\") pod \"ceilometer-0\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.715966 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.866003 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.969580 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76v4z\" (UniqueName: \"kubernetes.io/projected/43129653-d1cd-4ced-b120-33be343c3d9d-kube-api-access-76v4z\") pod \"43129653-d1cd-4ced-b120-33be343c3d9d\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.969756 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-combined-ca-bundle\") pod \"43129653-d1cd-4ced-b120-33be343c3d9d\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.969842 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-scripts\") pod \"43129653-d1cd-4ced-b120-33be343c3d9d\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.970092 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-config-data\") pod \"43129653-d1cd-4ced-b120-33be343c3d9d\" (UID: \"43129653-d1cd-4ced-b120-33be343c3d9d\") " Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.989401 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43129653-d1cd-4ced-b120-33be343c3d9d-kube-api-access-76v4z" (OuterVolumeSpecName: "kube-api-access-76v4z") pod "43129653-d1cd-4ced-b120-33be343c3d9d" (UID: "43129653-d1cd-4ced-b120-33be343c3d9d"). InnerVolumeSpecName "kube-api-access-76v4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:11 crc kubenswrapper[4690]: I1203 13:35:11.990155 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-scripts" (OuterVolumeSpecName: "scripts") pod "43129653-d1cd-4ced-b120-33be343c3d9d" (UID: "43129653-d1cd-4ced-b120-33be343c3d9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.021327 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-config-data" (OuterVolumeSpecName: "config-data") pod "43129653-d1cd-4ced-b120-33be343c3d9d" (UID: "43129653-d1cd-4ced-b120-33be343c3d9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.033717 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43129653-d1cd-4ced-b120-33be343c3d9d" (UID: "43129653-d1cd-4ced-b120-33be343c3d9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.072499 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.072529 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.072541 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43129653-d1cd-4ced-b120-33be343c3d9d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.072550 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76v4z\" (UniqueName: \"kubernetes.io/projected/43129653-d1cd-4ced-b120-33be343c3d9d-kube-api-access-76v4z\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.276494 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.316490 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.345391 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f4cdc2-0994-4a59-9a7a-c6680f10d359" path="/var/lib/kubelet/pods/d3f4cdc2-0994-4a59-9a7a-c6680f10d359/volumes" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.350858 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hvqqp" event={"ID":"43129653-d1cd-4ced-b120-33be343c3d9d","Type":"ContainerDied","Data":"619b74c0587fed892cf2a7148e4b2e50c5604961ceba5724e4b18d004bcb8770"} Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.352511 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="619b74c0587fed892cf2a7148e4b2e50c5604961ceba5724e4b18d004bcb8770" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.352534 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerStarted","Data":"892b432559168d0b6eaf6bff63e4958b39492ac73b720de3ee48d206976f8e14"} Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.435148 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 13:35:12 crc kubenswrapper[4690]: E1203 13:35:12.435664 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43129653-d1cd-4ced-b120-33be343c3d9d" containerName="nova-cell0-conductor-db-sync" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.435683 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="43129653-d1cd-4ced-b120-33be343c3d9d" containerName="nova-cell0-conductor-db-sync" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.435941 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="43129653-d1cd-4ced-b120-33be343c3d9d" containerName="nova-cell0-conductor-db-sync" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.436724 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.439098 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-l58jm" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.440133 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.454677 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.486604 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqzbg\" (UniqueName: \"kubernetes.io/projected/a730fe99-3893-4697-b946-f58470a93407-kube-api-access-wqzbg\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.486720 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a730fe99-3893-4697-b946-f58470a93407-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.486741 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a730fe99-3893-4697-b946-f58470a93407-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.588512 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqzbg\" (UniqueName: \"kubernetes.io/projected/a730fe99-3893-4697-b946-f58470a93407-kube-api-access-wqzbg\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.588736 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a730fe99-3893-4697-b946-f58470a93407-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.588775 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a730fe99-3893-4697-b946-f58470a93407-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.595028 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a730fe99-3893-4697-b946-f58470a93407-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.595050 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a730fe99-3893-4697-b946-f58470a93407-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.610355 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqzbg\" (UniqueName: \"kubernetes.io/projected/a730fe99-3893-4697-b946-f58470a93407-kube-api-access-wqzbg\") pod \"nova-cell0-conductor-0\" (UID: \"a730fe99-3893-4697-b946-f58470a93407\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:12 crc kubenswrapper[4690]: I1203 13:35:12.776231 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.334743 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerStarted","Data":"842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397"} Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.471608 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zlqww"] Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.484964 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.490529 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zlqww"] Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.515092 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-utilities\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.515175 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjlcq\" (UniqueName: \"kubernetes.io/projected/f5170fc3-75e2-401a-871d-1a1b50a5e88a-kube-api-access-fjlcq\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.515317 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-catalog-content\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.535878 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.618678 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-catalog-content\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.619048 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-utilities\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.619093 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjlcq\" (UniqueName: \"kubernetes.io/projected/f5170fc3-75e2-401a-871d-1a1b50a5e88a-kube-api-access-fjlcq\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.619496 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-catalog-content\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.619725 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-utilities\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.640596 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjlcq\" (UniqueName: \"kubernetes.io/projected/f5170fc3-75e2-401a-871d-1a1b50a5e88a-kube-api-access-fjlcq\") pod \"certified-operators-zlqww\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:13 crc kubenswrapper[4690]: I1203 13:35:13.813771 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:14 crc kubenswrapper[4690]: I1203 13:35:14.374807 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a730fe99-3893-4697-b946-f58470a93407","Type":"ContainerStarted","Data":"456728157f1974ef7d87189daba30d5dffde646101ae04b5059b039e770a9765"} Dec 03 13:35:14 crc kubenswrapper[4690]: I1203 13:35:14.375585 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a730fe99-3893-4697-b946-f58470a93407","Type":"ContainerStarted","Data":"57cc8312ae9ee31d259c40d314922075606c201f733869e31f54a8347552ef9e"} Dec 03 13:35:14 crc kubenswrapper[4690]: I1203 13:35:14.375708 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:14 crc kubenswrapper[4690]: I1203 13:35:14.491599 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zlqww"] Dec 03 13:35:14 crc kubenswrapper[4690]: I1203 13:35:14.542657 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.542624621 podStartE2EDuration="2.542624621s" podCreationTimestamp="2025-12-03 13:35:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:35:14.472715097 +0000 UTC m=+1560.453635540" watchObservedRunningTime="2025-12-03 13:35:14.542624621 +0000 UTC m=+1560.523545054" Dec 03 13:35:15 crc kubenswrapper[4690]: I1203 13:35:15.394890 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerStarted","Data":"e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff"} Dec 03 13:35:15 crc kubenswrapper[4690]: I1203 13:35:15.397645 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlqww" event={"ID":"f5170fc3-75e2-401a-871d-1a1b50a5e88a","Type":"ContainerStarted","Data":"1f5ed32ff03e58d3ebbe0b4b0a7f1bb2f0d6669184d51669025509c6b4f71e2f"} Dec 03 13:35:15 crc kubenswrapper[4690]: I1203 13:35:15.397683 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlqww" event={"ID":"f5170fc3-75e2-401a-871d-1a1b50a5e88a","Type":"ContainerStarted","Data":"2012220e5dbffd9ed9ba93a68a56e3bd0db082620b0f2d45c56a3d11594da859"} Dec 03 13:35:16 crc kubenswrapper[4690]: I1203 13:35:16.411944 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerStarted","Data":"8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8"} Dec 03 13:35:16 crc kubenswrapper[4690]: I1203 13:35:16.414623 4690 generic.go:334] "Generic (PLEG): container finished" podID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerID="1f5ed32ff03e58d3ebbe0b4b0a7f1bb2f0d6669184d51669025509c6b4f71e2f" exitCode=0 Dec 03 13:35:16 crc kubenswrapper[4690]: I1203 13:35:16.414688 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlqww" event={"ID":"f5170fc3-75e2-401a-871d-1a1b50a5e88a","Type":"ContainerDied","Data":"1f5ed32ff03e58d3ebbe0b4b0a7f1bb2f0d6669184d51669025509c6b4f71e2f"} Dec 03 13:35:18 crc kubenswrapper[4690]: I1203 13:35:18.453726 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerStarted","Data":"62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559"} Dec 03 13:35:18 crc kubenswrapper[4690]: I1203 13:35:18.454370 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:35:18 crc kubenswrapper[4690]: I1203 13:35:18.459779 4690 generic.go:334] "Generic (PLEG): container finished" podID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerID="a949414ada885fe75fb6f5df2fe1ee0298a0d0941be95c142bb92142dfbda839" exitCode=0 Dec 03 13:35:18 crc kubenswrapper[4690]: I1203 13:35:18.459822 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlqww" event={"ID":"f5170fc3-75e2-401a-871d-1a1b50a5e88a","Type":"ContainerDied","Data":"a949414ada885fe75fb6f5df2fe1ee0298a0d0941be95c142bb92142dfbda839"} Dec 03 13:35:18 crc kubenswrapper[4690]: I1203 13:35:18.488400 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.438456905 podStartE2EDuration="7.48836913s" podCreationTimestamp="2025-12-03 13:35:11 +0000 UTC" firstStartedPulling="2025-12-03 13:35:12.293476284 +0000 UTC m=+1558.274396717" lastFinishedPulling="2025-12-03 13:35:17.343388509 +0000 UTC m=+1563.324308942" observedRunningTime="2025-12-03 13:35:18.481148189 +0000 UTC m=+1564.462068622" watchObservedRunningTime="2025-12-03 13:35:18.48836913 +0000 UTC m=+1564.469289563" Dec 03 13:35:19 crc kubenswrapper[4690]: I1203 13:35:19.472516 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlqww" event={"ID":"f5170fc3-75e2-401a-871d-1a1b50a5e88a","Type":"ContainerStarted","Data":"f6550f31817b61da9d8a15363a53e2b8a4a880a35e5220b7dc13e28f18ce16f4"} Dec 03 13:35:19 crc kubenswrapper[4690]: I1203 13:35:19.507509 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zlqww" podStartSLOduration=3.896310225 podStartE2EDuration="6.507490052s" podCreationTimestamp="2025-12-03 13:35:13 +0000 UTC" firstStartedPulling="2025-12-03 13:35:16.416990512 +0000 UTC m=+1562.397910945" lastFinishedPulling="2025-12-03 13:35:19.028170339 +0000 UTC m=+1565.009090772" observedRunningTime="2025-12-03 13:35:19.489179769 +0000 UTC m=+1565.470100222" watchObservedRunningTime="2025-12-03 13:35:19.507490052 +0000 UTC m=+1565.488410485" Dec 03 13:35:22 crc kubenswrapper[4690]: I1203 13:35:22.808654 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.325640 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-2smlf"] Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.327107 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.331111 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.331261 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.345641 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2smlf"] Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.404595 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.404703 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-scripts\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.404771 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-config-data\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.404824 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4szs9\" (UniqueName: \"kubernetes.io/projected/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-kube-api-access-4szs9\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.506658 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.506727 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-scripts\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.506770 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-config-data\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.506829 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4szs9\" (UniqueName: \"kubernetes.io/projected/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-kube-api-access-4szs9\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.513664 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.514345 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-scripts\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.517831 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-config-data\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.524464 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4szs9\" (UniqueName: \"kubernetes.io/projected/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-kube-api-access-4szs9\") pod \"nova-cell0-cell-mapping-2smlf\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.651612 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.768028 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.770295 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.772353 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.788671 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.815017 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.816851 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.818470 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.818568 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-config-data\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.818637 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxr82\" (UniqueName: \"kubernetes.io/projected/c736f164-3d32-432d-9f36-ab2af39527b9-kube-api-access-qxr82\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.818726 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.818968 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.819025 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c736f164-3d32-432d-9f36-ab2af39527b9-logs\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.822085 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.838800 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.900524 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.922732 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c736f164-3d32-432d-9f36-ab2af39527b9-logs\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.923202 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-config-data\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.923365 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxr82\" (UniqueName: \"kubernetes.io/projected/c736f164-3d32-432d-9f36-ab2af39527b9-kube-api-access-qxr82\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.923484 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.923624 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.923824 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdmcm\" (UniqueName: \"kubernetes.io/projected/281194a0-06f2-460f-837d-4d27acce14f4-kube-api-access-wdmcm\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.923919 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-config-data\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.926396 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.928508 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c736f164-3d32-432d-9f36-ab2af39527b9-logs\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.928773 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.939514 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.947047 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-config-data\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.970525 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:23 crc kubenswrapper[4690]: I1203 13:35:23.989083 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.014576 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxr82\" (UniqueName: \"kubernetes.io/projected/c736f164-3d32-432d-9f36-ab2af39527b9-kube-api-access-qxr82\") pod \"nova-api-0\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " pod="openstack/nova-api-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.025949 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24bebacb-2b5e-45a7-8fb2-838d4d40193e-logs\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.026033 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdmcm\" (UniqueName: \"kubernetes.io/projected/281194a0-06f2-460f-837d-4d27acce14f4-kube-api-access-wdmcm\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.026062 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-config-data\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.026149 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjdbr\" (UniqueName: \"kubernetes.io/projected/24bebacb-2b5e-45a7-8fb2-838d4d40193e-kube-api-access-fjdbr\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.026202 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.026230 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.026316 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-config-data\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.071933 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-config-data\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.116737 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.117601 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.120150 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.122532 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdmcm\" (UniqueName: \"kubernetes.io/projected/281194a0-06f2-460f-837d-4d27acce14f4-kube-api-access-wdmcm\") pod \"nova-scheduler-0\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.122983 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.129570 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/7ce6e589-c9fb-40ec-9338-cfe095c1b158-kube-api-access-89vlq\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.129681 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.129760 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjdbr\" (UniqueName: \"kubernetes.io/projected/24bebacb-2b5e-45a7-8fb2-838d4d40193e-kube-api-access-fjdbr\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.129836 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.129992 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-config-data\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.130037 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.130067 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24bebacb-2b5e-45a7-8fb2-838d4d40193e-logs\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.130549 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24bebacb-2b5e-45a7-8fb2-838d4d40193e-logs\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.135061 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.136489 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-config-data\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.156925 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.167626 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.176576 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.177509 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjdbr\" (UniqueName: \"kubernetes.io/projected/24bebacb-2b5e-45a7-8fb2-838d4d40193e-kube-api-access-fjdbr\") pod \"nova-metadata-0\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.233116 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.233218 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/7ce6e589-c9fb-40ec-9338-cfe095c1b158-kube-api-access-89vlq\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.233338 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.243550 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.245788 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x5ws6"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.249232 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.257003 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.265736 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x5ws6"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.277710 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/7ce6e589-c9fb-40ec-9338-cfe095c1b158-kube-api-access-89vlq\") pod \"nova-cell1-novncproxy-0\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.356097 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4xnz\" (UniqueName: \"kubernetes.io/projected/7290142a-86a6-45d7-8f74-98fb5bf783ed-kube-api-access-q4xnz\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.356538 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.356582 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.356660 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.356723 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-config\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.356783 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.374989 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.456029 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.459847 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.460092 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4xnz\" (UniqueName: \"kubernetes.io/projected/7290142a-86a6-45d7-8f74-98fb5bf783ed-kube-api-access-q4xnz\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.460231 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.460261 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.460304 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.460352 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-config\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.461530 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-config\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.462207 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.461531 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.462319 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.464021 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.481001 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4xnz\" (UniqueName: \"kubernetes.io/projected/7290142a-86a6-45d7-8f74-98fb5bf783ed-kube-api-access-q4xnz\") pod \"dnsmasq-dns-757b4f8459-x5ws6\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.560256 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2smlf"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.615988 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.761378 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jgxms"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.765843 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.772433 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jgxms"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.773350 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.774148 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.822949 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.878695 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-config-data\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.879713 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.880248 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-scripts\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.880540 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9n75\" (UniqueName: \"kubernetes.io/projected/76a65eb6-765c-4360-bed0-c06558a1aad6-kube-api-access-j9n75\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.926609 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zlqww"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.958015 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.991373 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.991484 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-scripts\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.991575 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9n75\" (UniqueName: \"kubernetes.io/projected/76a65eb6-765c-4360-bed0-c06558a1aad6-kube-api-access-j9n75\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:24 crc kubenswrapper[4690]: I1203 13:35:24.991642 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-config-data\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.000836 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-scripts\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.001289 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.002706 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-config-data\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.017424 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9n75\" (UniqueName: \"kubernetes.io/projected/76a65eb6-765c-4360-bed0-c06558a1aad6-kube-api-access-j9n75\") pod \"nova-cell1-conductor-db-sync-jgxms\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.153828 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.473555 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.488155 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.618434 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.681550 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2smlf" event={"ID":"d4ae925d-b15a-411f-b2b8-64ddbbbdb097","Type":"ContainerStarted","Data":"83d3b49c42eccb6643a650c91780e185a7df0744daad78fb2761714a83574e44"} Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.681611 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2smlf" event={"ID":"d4ae925d-b15a-411f-b2b8-64ddbbbdb097","Type":"ContainerStarted","Data":"622091476daacde0abe81bcc32ef66b15f9d4cc8012ff5a3d68bcc99c4633046"} Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.684516 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7ce6e589-c9fb-40ec-9338-cfe095c1b158","Type":"ContainerStarted","Data":"5b90349fb98269c97361d16508254becba97c6eebbe2944b9709c0f339dbe43c"} Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.687257 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c736f164-3d32-432d-9f36-ab2af39527b9","Type":"ContainerStarted","Data":"83f2d47ee9689278b1291fcd8a2c2310a0501f7a4c9cd783bfac08745961cfbb"} Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.688998 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24bebacb-2b5e-45a7-8fb2-838d4d40193e","Type":"ContainerStarted","Data":"b3c39ff26b3ee64c59aa4eee3578760b896f001eb98c25f2383ea9d1f3b9f7d4"} Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.705262 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"281194a0-06f2-460f-837d-4d27acce14f4","Type":"ContainerStarted","Data":"31997bbad68cbd62fb2e21e45683e599b7d74267c09fc48ed87bdb8493c4ca47"} Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.709710 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-2smlf" podStartSLOduration=2.709684251 podStartE2EDuration="2.709684251s" podCreationTimestamp="2025-12-03 13:35:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:35:25.69902128 +0000 UTC m=+1571.679941723" watchObservedRunningTime="2025-12-03 13:35:25.709684251 +0000 UTC m=+1571.690604684" Dec 03 13:35:25 crc kubenswrapper[4690]: I1203 13:35:25.871646 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x5ws6"] Dec 03 13:35:25 crc kubenswrapper[4690]: W1203 13:35:25.872966 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7290142a_86a6_45d7_8f74_98fb5bf783ed.slice/crio-1ce909fe3b2ad7e748b5eaa9bf20d1f0bcc641dcadbaa832c56a7e4682e5f009 WatchSource:0}: Error finding container 1ce909fe3b2ad7e748b5eaa9bf20d1f0bcc641dcadbaa832c56a7e4682e5f009: Status 404 returned error can't find the container with id 1ce909fe3b2ad7e748b5eaa9bf20d1f0bcc641dcadbaa832c56a7e4682e5f009 Dec 03 13:35:26 crc kubenswrapper[4690]: I1203 13:35:26.070072 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jgxms"] Dec 03 13:35:26 crc kubenswrapper[4690]: I1203 13:35:26.749604 4690 generic.go:334] "Generic (PLEG): container finished" podID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerID="99a0b9129c7bb324033787adefc3f16a12ea5f107eaa7cb0d938bce154d446ef" exitCode=0 Dec 03 13:35:26 crc kubenswrapper[4690]: I1203 13:35:26.750471 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" event={"ID":"7290142a-86a6-45d7-8f74-98fb5bf783ed","Type":"ContainerDied","Data":"99a0b9129c7bb324033787adefc3f16a12ea5f107eaa7cb0d938bce154d446ef"} Dec 03 13:35:26 crc kubenswrapper[4690]: I1203 13:35:26.750540 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" event={"ID":"7290142a-86a6-45d7-8f74-98fb5bf783ed","Type":"ContainerStarted","Data":"1ce909fe3b2ad7e748b5eaa9bf20d1f0bcc641dcadbaa832c56a7e4682e5f009"} Dec 03 13:35:26 crc kubenswrapper[4690]: I1203 13:35:26.757113 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zlqww" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerName="registry-server" containerID="cri-o://f6550f31817b61da9d8a15363a53e2b8a4a880a35e5220b7dc13e28f18ce16f4" gracePeriod=2 Dec 03 13:35:26 crc kubenswrapper[4690]: I1203 13:35:26.757386 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jgxms" event={"ID":"76a65eb6-765c-4360-bed0-c06558a1aad6","Type":"ContainerStarted","Data":"b6c576df6574c4663931f9eec991081ae18b6f741b1655bb6916370a999de9bd"} Dec 03 13:35:26 crc kubenswrapper[4690]: I1203 13:35:26.757623 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jgxms" event={"ID":"76a65eb6-765c-4360-bed0-c06558a1aad6","Type":"ContainerStarted","Data":"24aed8af298f9b2d89b8823aef4b291296a5fb6ff0b2698b653874b91a7c6d1b"} Dec 03 13:35:26 crc kubenswrapper[4690]: I1203 13:35:26.850055 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-jgxms" podStartSLOduration=2.850030771 podStartE2EDuration="2.850030771s" podCreationTimestamp="2025-12-03 13:35:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:35:26.805481866 +0000 UTC m=+1572.786402299" watchObservedRunningTime="2025-12-03 13:35:26.850030771 +0000 UTC m=+1572.830951204" Dec 03 13:35:27 crc kubenswrapper[4690]: I1203 13:35:27.333927 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:35:27 crc kubenswrapper[4690]: I1203 13:35:27.415063 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:35:27 crc kubenswrapper[4690]: I1203 13:35:27.810461 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" event={"ID":"7290142a-86a6-45d7-8f74-98fb5bf783ed","Type":"ContainerStarted","Data":"d57dfb990b6a701faa8a781c512958741ff80692d4ce3558add0eaff454a2a27"} Dec 03 13:35:27 crc kubenswrapper[4690]: I1203 13:35:27.810784 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:27 crc kubenswrapper[4690]: I1203 13:35:27.827137 4690 generic.go:334] "Generic (PLEG): container finished" podID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerID="f6550f31817b61da9d8a15363a53e2b8a4a880a35e5220b7dc13e28f18ce16f4" exitCode=0 Dec 03 13:35:27 crc kubenswrapper[4690]: I1203 13:35:27.827776 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlqww" event={"ID":"f5170fc3-75e2-401a-871d-1a1b50a5e88a","Type":"ContainerDied","Data":"f6550f31817b61da9d8a15363a53e2b8a4a880a35e5220b7dc13e28f18ce16f4"} Dec 03 13:35:27 crc kubenswrapper[4690]: I1203 13:35:27.875209 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" podStartSLOduration=3.875183882 podStartE2EDuration="3.875183882s" podCreationTimestamp="2025-12-03 13:35:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:35:27.868983859 +0000 UTC m=+1573.849904292" watchObservedRunningTime="2025-12-03 13:35:27.875183882 +0000 UTC m=+1573.856104315" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.782940 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.851693 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlqww" event={"ID":"f5170fc3-75e2-401a-871d-1a1b50a5e88a","Type":"ContainerDied","Data":"2012220e5dbffd9ed9ba93a68a56e3bd0db082620b0f2d45c56a3d11594da859"} Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.851734 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlqww" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.851778 4690 scope.go:117] "RemoveContainer" containerID="f6550f31817b61da9d8a15363a53e2b8a4a880a35e5220b7dc13e28f18ce16f4" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.856919 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-utilities\") pod \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.857045 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjlcq\" (UniqueName: \"kubernetes.io/projected/f5170fc3-75e2-401a-871d-1a1b50a5e88a-kube-api-access-fjlcq\") pod \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.857262 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-catalog-content\") pod \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\" (UID: \"f5170fc3-75e2-401a-871d-1a1b50a5e88a\") " Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.858019 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-utilities" (OuterVolumeSpecName: "utilities") pod "f5170fc3-75e2-401a-871d-1a1b50a5e88a" (UID: "f5170fc3-75e2-401a-871d-1a1b50a5e88a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.870594 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5170fc3-75e2-401a-871d-1a1b50a5e88a-kube-api-access-fjlcq" (OuterVolumeSpecName: "kube-api-access-fjlcq") pod "f5170fc3-75e2-401a-871d-1a1b50a5e88a" (UID: "f5170fc3-75e2-401a-871d-1a1b50a5e88a"). InnerVolumeSpecName "kube-api-access-fjlcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.906049 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5170fc3-75e2-401a-871d-1a1b50a5e88a" (UID: "f5170fc3-75e2-401a-871d-1a1b50a5e88a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.959532 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.959571 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjlcq\" (UniqueName: \"kubernetes.io/projected/f5170fc3-75e2-401a-871d-1a1b50a5e88a-kube-api-access-fjlcq\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:28 crc kubenswrapper[4690]: I1203 13:35:28.959587 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5170fc3-75e2-401a-871d-1a1b50a5e88a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:29 crc kubenswrapper[4690]: I1203 13:35:29.197882 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zlqww"] Dec 03 13:35:29 crc kubenswrapper[4690]: I1203 13:35:29.207933 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zlqww"] Dec 03 13:35:29 crc kubenswrapper[4690]: I1203 13:35:29.727938 4690 scope.go:117] "RemoveContainer" containerID="a949414ada885fe75fb6f5df2fe1ee0298a0d0941be95c142bb92142dfbda839" Dec 03 13:35:29 crc kubenswrapper[4690]: I1203 13:35:29.789375 4690 scope.go:117] "RemoveContainer" containerID="1f5ed32ff03e58d3ebbe0b4b0a7f1bb2f0d6669184d51669025509c6b4f71e2f" Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.328576 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" path="/var/lib/kubelet/pods/f5170fc3-75e2-401a-871d-1a1b50a5e88a/volumes" Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.888670 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24bebacb-2b5e-45a7-8fb2-838d4d40193e","Type":"ContainerStarted","Data":"c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687"} Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.888737 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24bebacb-2b5e-45a7-8fb2-838d4d40193e","Type":"ContainerStarted","Data":"8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916"} Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.888808 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerName="nova-metadata-log" containerID="cri-o://8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916" gracePeriod=30 Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.888837 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerName="nova-metadata-metadata" containerID="cri-o://c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687" gracePeriod=30 Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.894803 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"281194a0-06f2-460f-837d-4d27acce14f4","Type":"ContainerStarted","Data":"bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1"} Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.908768 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7ce6e589-c9fb-40ec-9338-cfe095c1b158","Type":"ContainerStarted","Data":"d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6"} Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.908939 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="7ce6e589-c9fb-40ec-9338-cfe095c1b158" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6" gracePeriod=30 Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.922217 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c736f164-3d32-432d-9f36-ab2af39527b9","Type":"ContainerStarted","Data":"63d55a067ec1a6c03055428043e3f1abbda4194b04511e773d4776774a49d9f4"} Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.922279 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c736f164-3d32-432d-9f36-ab2af39527b9","Type":"ContainerStarted","Data":"dd74e69f776ba57e0d316bf3a884031e82f948863c6882f0d80a0d5844a2fa6d"} Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.925167 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.612939067 podStartE2EDuration="7.925145963s" podCreationTimestamp="2025-12-03 13:35:23 +0000 UTC" firstStartedPulling="2025-12-03 13:35:25.495015828 +0000 UTC m=+1571.475936261" lastFinishedPulling="2025-12-03 13:35:29.807222734 +0000 UTC m=+1575.788143157" observedRunningTime="2025-12-03 13:35:30.908072043 +0000 UTC m=+1576.888992486" watchObservedRunningTime="2025-12-03 13:35:30.925145963 +0000 UTC m=+1576.906066386" Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.954042 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.649015998 podStartE2EDuration="7.954012964s" podCreationTimestamp="2025-12-03 13:35:23 +0000 UTC" firstStartedPulling="2025-12-03 13:35:25.494609858 +0000 UTC m=+1571.475530291" lastFinishedPulling="2025-12-03 13:35:29.799606824 +0000 UTC m=+1575.780527257" observedRunningTime="2025-12-03 13:35:30.936808301 +0000 UTC m=+1576.917728724" watchObservedRunningTime="2025-12-03 13:35:30.954012964 +0000 UTC m=+1576.934933397" Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.963238 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.768945122 podStartE2EDuration="7.963210757s" podCreationTimestamp="2025-12-03 13:35:23 +0000 UTC" firstStartedPulling="2025-12-03 13:35:25.622402149 +0000 UTC m=+1571.603322582" lastFinishedPulling="2025-12-03 13:35:29.816667784 +0000 UTC m=+1575.797588217" observedRunningTime="2025-12-03 13:35:30.952757011 +0000 UTC m=+1576.933677444" watchObservedRunningTime="2025-12-03 13:35:30.963210757 +0000 UTC m=+1576.944131190" Dec 03 13:35:30 crc kubenswrapper[4690]: I1203 13:35:30.984071 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.185432982 podStartE2EDuration="7.984052317s" podCreationTimestamp="2025-12-03 13:35:23 +0000 UTC" firstStartedPulling="2025-12-03 13:35:24.970904685 +0000 UTC m=+1570.951825118" lastFinishedPulling="2025-12-03 13:35:29.76952402 +0000 UTC m=+1575.750444453" observedRunningTime="2025-12-03 13:35:30.974558356 +0000 UTC m=+1576.955478789" watchObservedRunningTime="2025-12-03 13:35:30.984052317 +0000 UTC m=+1576.964972750" Dec 03 13:35:31 crc kubenswrapper[4690]: I1203 13:35:31.935046 4690 generic.go:334] "Generic (PLEG): container finished" podID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerID="8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916" exitCode=143 Dec 03 13:35:31 crc kubenswrapper[4690]: I1203 13:35:31.935170 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24bebacb-2b5e-45a7-8fb2-838d4d40193e","Type":"ContainerDied","Data":"8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916"} Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.159105 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.159530 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.169120 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.169173 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.198110 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.376124 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.376220 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.456742 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.622101 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.746288 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-sw2bh"] Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.746775 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" podUID="e70496e1-182e-4071-9170-ab8bf54a14ee" containerName="dnsmasq-dns" containerID="cri-o://192f75d3c726653dccdb15141796efc6d4b6c451829082723fd8d74f18e4583b" gracePeriod=10 Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.996916 4690 generic.go:334] "Generic (PLEG): container finished" podID="e70496e1-182e-4071-9170-ab8bf54a14ee" containerID="192f75d3c726653dccdb15141796efc6d4b6c451829082723fd8d74f18e4583b" exitCode=0 Dec 03 13:35:34 crc kubenswrapper[4690]: I1203 13:35:34.997532 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" event={"ID":"e70496e1-182e-4071-9170-ab8bf54a14ee","Type":"ContainerDied","Data":"192f75d3c726653dccdb15141796efc6d4b6c451829082723fd8d74f18e4583b"} Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.007733 4690 generic.go:334] "Generic (PLEG): container finished" podID="d4ae925d-b15a-411f-b2b8-64ddbbbdb097" containerID="83d3b49c42eccb6643a650c91780e185a7df0744daad78fb2761714a83574e44" exitCode=0 Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.008622 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2smlf" event={"ID":"d4ae925d-b15a-411f-b2b8-64ddbbbdb097","Type":"ContainerDied","Data":"83d3b49c42eccb6643a650c91780e185a7df0744daad78fb2761714a83574e44"} Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.056817 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 13:35:35 crc kubenswrapper[4690]: E1203 13:35:35.092960 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4ae925d_b15a_411f_b2b8_64ddbbbdb097.slice/crio-83d3b49c42eccb6643a650c91780e185a7df0744daad78fb2761714a83574e44.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4ae925d_b15a_411f_b2b8_64ddbbbdb097.slice/crio-conmon-83d3b49c42eccb6643a650c91780e185a7df0744daad78fb2761714a83574e44.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.211823 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.253064 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.503658 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.556209 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b45q6\" (UniqueName: \"kubernetes.io/projected/e70496e1-182e-4071-9170-ab8bf54a14ee-kube-api-access-b45q6\") pod \"e70496e1-182e-4071-9170-ab8bf54a14ee\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.556342 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-swift-storage-0\") pod \"e70496e1-182e-4071-9170-ab8bf54a14ee\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.556366 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-svc\") pod \"e70496e1-182e-4071-9170-ab8bf54a14ee\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.556796 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-config\") pod \"e70496e1-182e-4071-9170-ab8bf54a14ee\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.556963 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-nb\") pod \"e70496e1-182e-4071-9170-ab8bf54a14ee\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.557033 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-sb\") pod \"e70496e1-182e-4071-9170-ab8bf54a14ee\" (UID: \"e70496e1-182e-4071-9170-ab8bf54a14ee\") " Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.567953 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70496e1-182e-4071-9170-ab8bf54a14ee-kube-api-access-b45q6" (OuterVolumeSpecName: "kube-api-access-b45q6") pod "e70496e1-182e-4071-9170-ab8bf54a14ee" (UID: "e70496e1-182e-4071-9170-ab8bf54a14ee"). InnerVolumeSpecName "kube-api-access-b45q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.660060 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b45q6\" (UniqueName: \"kubernetes.io/projected/e70496e1-182e-4071-9170-ab8bf54a14ee-kube-api-access-b45q6\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.666628 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e70496e1-182e-4071-9170-ab8bf54a14ee" (UID: "e70496e1-182e-4071-9170-ab8bf54a14ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.675140 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e70496e1-182e-4071-9170-ab8bf54a14ee" (UID: "e70496e1-182e-4071-9170-ab8bf54a14ee"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.680806 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e70496e1-182e-4071-9170-ab8bf54a14ee" (UID: "e70496e1-182e-4071-9170-ab8bf54a14ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.682301 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e70496e1-182e-4071-9170-ab8bf54a14ee" (UID: "e70496e1-182e-4071-9170-ab8bf54a14ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.697642 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-config" (OuterVolumeSpecName: "config") pod "e70496e1-182e-4071-9170-ab8bf54a14ee" (UID: "e70496e1-182e-4071-9170-ab8bf54a14ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.763124 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.763719 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.763838 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.763962 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:35 crc kubenswrapper[4690]: I1203 13:35:35.764055 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e70496e1-182e-4071-9170-ab8bf54a14ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.022711 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.022724 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-sw2bh" event={"ID":"e70496e1-182e-4071-9170-ab8bf54a14ee","Type":"ContainerDied","Data":"4261caa0df63055f253562529221cbcf76e2264af0c1f34abf171cdeff076f6b"} Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.023245 4690 scope.go:117] "RemoveContainer" containerID="192f75d3c726653dccdb15141796efc6d4b6c451829082723fd8d74f18e4583b" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.091232 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-sw2bh"] Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.101256 4690 scope.go:117] "RemoveContainer" containerID="c9385cdab24a1523c9f7967aaccad46c7f31abf218fc1aff032867ebf7cf3bde" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.104482 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-sw2bh"] Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.332575 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e70496e1-182e-4071-9170-ab8bf54a14ee" path="/var/lib/kubelet/pods/e70496e1-182e-4071-9170-ab8bf54a14ee/volumes" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.461167 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.591551 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4szs9\" (UniqueName: \"kubernetes.io/projected/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-kube-api-access-4szs9\") pod \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.591751 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-config-data\") pod \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.591788 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-scripts\") pod \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.591838 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-combined-ca-bundle\") pod \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\" (UID: \"d4ae925d-b15a-411f-b2b8-64ddbbbdb097\") " Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.598692 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-kube-api-access-4szs9" (OuterVolumeSpecName: "kube-api-access-4szs9") pod "d4ae925d-b15a-411f-b2b8-64ddbbbdb097" (UID: "d4ae925d-b15a-411f-b2b8-64ddbbbdb097"). InnerVolumeSpecName "kube-api-access-4szs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.599159 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-scripts" (OuterVolumeSpecName: "scripts") pod "d4ae925d-b15a-411f-b2b8-64ddbbbdb097" (UID: "d4ae925d-b15a-411f-b2b8-64ddbbbdb097"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.627932 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-config-data" (OuterVolumeSpecName: "config-data") pod "d4ae925d-b15a-411f-b2b8-64ddbbbdb097" (UID: "d4ae925d-b15a-411f-b2b8-64ddbbbdb097"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.632604 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4ae925d-b15a-411f-b2b8-64ddbbbdb097" (UID: "d4ae925d-b15a-411f-b2b8-64ddbbbdb097"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.695210 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4szs9\" (UniqueName: \"kubernetes.io/projected/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-kube-api-access-4szs9\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.695250 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.695261 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:36 crc kubenswrapper[4690]: I1203 13:35:36.695269 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4ae925d-b15a-411f-b2b8-64ddbbbdb097-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:37 crc kubenswrapper[4690]: I1203 13:35:37.039488 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2smlf" event={"ID":"d4ae925d-b15a-411f-b2b8-64ddbbbdb097","Type":"ContainerDied","Data":"622091476daacde0abe81bcc32ef66b15f9d4cc8012ff5a3d68bcc99c4633046"} Dec 03 13:35:37 crc kubenswrapper[4690]: I1203 13:35:37.039785 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="622091476daacde0abe81bcc32ef66b15f9d4cc8012ff5a3d68bcc99c4633046" Dec 03 13:35:37 crc kubenswrapper[4690]: I1203 13:35:37.039735 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2smlf" Dec 03 13:35:37 crc kubenswrapper[4690]: I1203 13:35:37.236372 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:37 crc kubenswrapper[4690]: I1203 13:35:37.236728 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-log" containerID="cri-o://dd74e69f776ba57e0d316bf3a884031e82f948863c6882f0d80a0d5844a2fa6d" gracePeriod=30 Dec 03 13:35:37 crc kubenswrapper[4690]: I1203 13:35:37.236961 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-api" containerID="cri-o://63d55a067ec1a6c03055428043e3f1abbda4194b04511e773d4776774a49d9f4" gracePeriod=30 Dec 03 13:35:37 crc kubenswrapper[4690]: I1203 13:35:37.262492 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:37 crc kubenswrapper[4690]: I1203 13:35:37.262735 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="281194a0-06f2-460f-837d-4d27acce14f4" containerName="nova-scheduler-scheduler" containerID="cri-o://bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1" gracePeriod=30 Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.050713 4690 generic.go:334] "Generic (PLEG): container finished" podID="c736f164-3d32-432d-9f36-ab2af39527b9" containerID="dd74e69f776ba57e0d316bf3a884031e82f948863c6882f0d80a0d5844a2fa6d" exitCode=143 Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.050829 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c736f164-3d32-432d-9f36-ab2af39527b9","Type":"ContainerDied","Data":"dd74e69f776ba57e0d316bf3a884031e82f948863c6882f0d80a0d5844a2fa6d"} Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.053150 4690 generic.go:334] "Generic (PLEG): container finished" podID="76a65eb6-765c-4360-bed0-c06558a1aad6" containerID="b6c576df6574c4663931f9eec991081ae18b6f741b1655bb6916370a999de9bd" exitCode=0 Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.053193 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jgxms" event={"ID":"76a65eb6-765c-4360-bed0-c06558a1aad6","Type":"ContainerDied","Data":"b6c576df6574c4663931f9eec991081ae18b6f741b1655bb6916370a999de9bd"} Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.892774 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.950423 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-config-data\") pod \"281194a0-06f2-460f-837d-4d27acce14f4\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.950613 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-combined-ca-bundle\") pod \"281194a0-06f2-460f-837d-4d27acce14f4\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.950649 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdmcm\" (UniqueName: \"kubernetes.io/projected/281194a0-06f2-460f-837d-4d27acce14f4-kube-api-access-wdmcm\") pod \"281194a0-06f2-460f-837d-4d27acce14f4\" (UID: \"281194a0-06f2-460f-837d-4d27acce14f4\") " Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.960844 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/281194a0-06f2-460f-837d-4d27acce14f4-kube-api-access-wdmcm" (OuterVolumeSpecName: "kube-api-access-wdmcm") pod "281194a0-06f2-460f-837d-4d27acce14f4" (UID: "281194a0-06f2-460f-837d-4d27acce14f4"). InnerVolumeSpecName "kube-api-access-wdmcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.985104 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-config-data" (OuterVolumeSpecName: "config-data") pod "281194a0-06f2-460f-837d-4d27acce14f4" (UID: "281194a0-06f2-460f-837d-4d27acce14f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:38 crc kubenswrapper[4690]: I1203 13:35:38.993780 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "281194a0-06f2-460f-837d-4d27acce14f4" (UID: "281194a0-06f2-460f-837d-4d27acce14f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.053422 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.053480 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdmcm\" (UniqueName: \"kubernetes.io/projected/281194a0-06f2-460f-837d-4d27acce14f4-kube-api-access-wdmcm\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.053495 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/281194a0-06f2-460f-837d-4d27acce14f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.066255 4690 generic.go:334] "Generic (PLEG): container finished" podID="281194a0-06f2-460f-837d-4d27acce14f4" containerID="bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1" exitCode=0 Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.066320 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"281194a0-06f2-460f-837d-4d27acce14f4","Type":"ContainerDied","Data":"bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1"} Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.066390 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"281194a0-06f2-460f-837d-4d27acce14f4","Type":"ContainerDied","Data":"31997bbad68cbd62fb2e21e45683e599b7d74267c09fc48ed87bdb8493c4ca47"} Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.066416 4690 scope.go:117] "RemoveContainer" containerID="bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.066331 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.106549 4690 scope.go:117] "RemoveContainer" containerID="bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1" Dec 03 13:35:39 crc kubenswrapper[4690]: E1203 13:35:39.108415 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1\": container with ID starting with bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1 not found: ID does not exist" containerID="bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.108449 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1"} err="failed to get container status \"bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1\": rpc error: code = NotFound desc = could not find container \"bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1\": container with ID starting with bbbae56267c2a8bc6b8ae190ad50fe3cf161f78cf50879b54f9a650d04a6bfc1 not found: ID does not exist" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.126719 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.143346 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.153498 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:39 crc kubenswrapper[4690]: E1203 13:35:39.154276 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4ae925d-b15a-411f-b2b8-64ddbbbdb097" containerName="nova-manage" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154311 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4ae925d-b15a-411f-b2b8-64ddbbbdb097" containerName="nova-manage" Dec 03 13:35:39 crc kubenswrapper[4690]: E1203 13:35:39.154335 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70496e1-182e-4071-9170-ab8bf54a14ee" containerName="init" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154360 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70496e1-182e-4071-9170-ab8bf54a14ee" containerName="init" Dec 03 13:35:39 crc kubenswrapper[4690]: E1203 13:35:39.154378 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerName="extract-content" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154386 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerName="extract-content" Dec 03 13:35:39 crc kubenswrapper[4690]: E1203 13:35:39.154404 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70496e1-182e-4071-9170-ab8bf54a14ee" containerName="dnsmasq-dns" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154415 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70496e1-182e-4071-9170-ab8bf54a14ee" containerName="dnsmasq-dns" Dec 03 13:35:39 crc kubenswrapper[4690]: E1203 13:35:39.154436 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281194a0-06f2-460f-837d-4d27acce14f4" containerName="nova-scheduler-scheduler" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154467 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="281194a0-06f2-460f-837d-4d27acce14f4" containerName="nova-scheduler-scheduler" Dec 03 13:35:39 crc kubenswrapper[4690]: E1203 13:35:39.154484 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerName="extract-utilities" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154491 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerName="extract-utilities" Dec 03 13:35:39 crc kubenswrapper[4690]: E1203 13:35:39.154522 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerName="registry-server" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154528 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerName="registry-server" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154779 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="281194a0-06f2-460f-837d-4d27acce14f4" containerName="nova-scheduler-scheduler" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154804 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5170fc3-75e2-401a-871d-1a1b50a5e88a" containerName="registry-server" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154814 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4ae925d-b15a-411f-b2b8-64ddbbbdb097" containerName="nova-manage" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.154834 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70496e1-182e-4071-9170-ab8bf54a14ee" containerName="dnsmasq-dns" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.156263 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.159859 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.163675 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.258666 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.258758 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrpgr\" (UniqueName: \"kubernetes.io/projected/345cf534-66c9-430b-b919-25add004cfdc-kube-api-access-wrpgr\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.258999 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-config-data\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.361054 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.361590 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrpgr\" (UniqueName: \"kubernetes.io/projected/345cf534-66c9-430b-b919-25add004cfdc-kube-api-access-wrpgr\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.361924 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-config-data\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.371619 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-config-data\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.371629 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.392996 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrpgr\" (UniqueName: \"kubernetes.io/projected/345cf534-66c9-430b-b919-25add004cfdc-kube-api-access-wrpgr\") pod \"nova-scheduler-0\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.485584 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.595251 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.668102 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-combined-ca-bundle\") pod \"76a65eb6-765c-4360-bed0-c06558a1aad6\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.668582 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-config-data\") pod \"76a65eb6-765c-4360-bed0-c06558a1aad6\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.668652 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9n75\" (UniqueName: \"kubernetes.io/projected/76a65eb6-765c-4360-bed0-c06558a1aad6-kube-api-access-j9n75\") pod \"76a65eb6-765c-4360-bed0-c06558a1aad6\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.668926 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-scripts\") pod \"76a65eb6-765c-4360-bed0-c06558a1aad6\" (UID: \"76a65eb6-765c-4360-bed0-c06558a1aad6\") " Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.674123 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-scripts" (OuterVolumeSpecName: "scripts") pod "76a65eb6-765c-4360-bed0-c06558a1aad6" (UID: "76a65eb6-765c-4360-bed0-c06558a1aad6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.680618 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a65eb6-765c-4360-bed0-c06558a1aad6-kube-api-access-j9n75" (OuterVolumeSpecName: "kube-api-access-j9n75") pod "76a65eb6-765c-4360-bed0-c06558a1aad6" (UID: "76a65eb6-765c-4360-bed0-c06558a1aad6"). InnerVolumeSpecName "kube-api-access-j9n75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.707053 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76a65eb6-765c-4360-bed0-c06558a1aad6" (UID: "76a65eb6-765c-4360-bed0-c06558a1aad6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.716297 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-config-data" (OuterVolumeSpecName: "config-data") pod "76a65eb6-765c-4360-bed0-c06558a1aad6" (UID: "76a65eb6-765c-4360-bed0-c06558a1aad6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.773888 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.773940 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.773958 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a65eb6-765c-4360-bed0-c06558a1aad6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.773975 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9n75\" (UniqueName: \"kubernetes.io/projected/76a65eb6-765c-4360-bed0-c06558a1aad6-kube-api-access-j9n75\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:39 crc kubenswrapper[4690]: I1203 13:35:39.986374 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.080005 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jgxms" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.080062 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jgxms" event={"ID":"76a65eb6-765c-4360-bed0-c06558a1aad6","Type":"ContainerDied","Data":"24aed8af298f9b2d89b8823aef4b291296a5fb6ff0b2698b653874b91a7c6d1b"} Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.080108 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24aed8af298f9b2d89b8823aef4b291296a5fb6ff0b2698b653874b91a7c6d1b" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.081066 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"345cf534-66c9-430b-b919-25add004cfdc","Type":"ContainerStarted","Data":"0ab43219c13031c5ae8795210fa06f45f080ee62db1a6f959e154a6995ab3a2c"} Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.175978 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 13:35:40 crc kubenswrapper[4690]: E1203 13:35:40.176515 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a65eb6-765c-4360-bed0-c06558a1aad6" containerName="nova-cell1-conductor-db-sync" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.176539 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a65eb6-765c-4360-bed0-c06558a1aad6" containerName="nova-cell1-conductor-db-sync" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.176815 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a65eb6-765c-4360-bed0-c06558a1aad6" containerName="nova-cell1-conductor-db-sync" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.177605 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.188337 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.190420 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.282959 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlsqf\" (UniqueName: \"kubernetes.io/projected/3141a41c-66ea-41fe-bc21-dd115b3cb98e-kube-api-access-mlsqf\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.283063 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3141a41c-66ea-41fe-bc21-dd115b3cb98e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.283121 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3141a41c-66ea-41fe-bc21-dd115b3cb98e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.329889 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="281194a0-06f2-460f-837d-4d27acce14f4" path="/var/lib/kubelet/pods/281194a0-06f2-460f-837d-4d27acce14f4/volumes" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.384538 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlsqf\" (UniqueName: \"kubernetes.io/projected/3141a41c-66ea-41fe-bc21-dd115b3cb98e-kube-api-access-mlsqf\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.384761 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3141a41c-66ea-41fe-bc21-dd115b3cb98e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.384895 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3141a41c-66ea-41fe-bc21-dd115b3cb98e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.392583 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3141a41c-66ea-41fe-bc21-dd115b3cb98e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.394705 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3141a41c-66ea-41fe-bc21-dd115b3cb98e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.408545 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlsqf\" (UniqueName: \"kubernetes.io/projected/3141a41c-66ea-41fe-bc21-dd115b3cb98e-kube-api-access-mlsqf\") pod \"nova-cell1-conductor-0\" (UID: \"3141a41c-66ea-41fe-bc21-dd115b3cb98e\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:40 crc kubenswrapper[4690]: I1203 13:35:40.497939 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.029907 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.103860 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"345cf534-66c9-430b-b919-25add004cfdc","Type":"ContainerStarted","Data":"e2deefc31f7e41d9f1c361ab9de4f03d2af25682181758a2fd11d0ceaac50882"} Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.106449 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3141a41c-66ea-41fe-bc21-dd115b3cb98e","Type":"ContainerStarted","Data":"1520efffab7d8da134d7fbd7eb59cb427be6fdfe18c4d997259ce128ea1336ea"} Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.131217 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.131166985 podStartE2EDuration="2.131166985s" podCreationTimestamp="2025-12-03 13:35:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:35:41.124116639 +0000 UTC m=+1587.105037072" watchObservedRunningTime="2025-12-03 13:35:41.131166985 +0000 UTC m=+1587.112087418" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.609735 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jnfn6"] Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.613251 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.624274 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnfn6"] Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.718795 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdflw\" (UniqueName: \"kubernetes.io/projected/98f79021-b882-441a-afb1-52637dd48566-kube-api-access-wdflw\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.718956 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-catalog-content\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.719059 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-utilities\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.724919 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.822307 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdflw\" (UniqueName: \"kubernetes.io/projected/98f79021-b882-441a-afb1-52637dd48566-kube-api-access-wdflw\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.822456 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-catalog-content\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.822659 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-utilities\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.823493 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-utilities\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.825480 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-catalog-content\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.854733 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdflw\" (UniqueName: \"kubernetes.io/projected/98f79021-b882-441a-afb1-52637dd48566-kube-api-access-wdflw\") pod \"redhat-marketplace-jnfn6\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:41 crc kubenswrapper[4690]: I1203 13:35:41.942169 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.150513 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3141a41c-66ea-41fe-bc21-dd115b3cb98e","Type":"ContainerStarted","Data":"279a14b0b4bb77a2aa85dc4c103dfddcd7fda99ff59a6db14ac877d16e229294"} Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.151059 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.172432 4690 generic.go:334] "Generic (PLEG): container finished" podID="c736f164-3d32-432d-9f36-ab2af39527b9" containerID="63d55a067ec1a6c03055428043e3f1abbda4194b04511e773d4776774a49d9f4" exitCode=0 Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.173416 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c736f164-3d32-432d-9f36-ab2af39527b9","Type":"ContainerDied","Data":"63d55a067ec1a6c03055428043e3f1abbda4194b04511e773d4776774a49d9f4"} Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.178934 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.178917043 podStartE2EDuration="2.178917043s" podCreationTimestamp="2025-12-03 13:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:35:42.170665805 +0000 UTC m=+1588.151586238" watchObservedRunningTime="2025-12-03 13:35:42.178917043 +0000 UTC m=+1588.159837476" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.268826 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.345947 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-combined-ca-bundle\") pod \"c736f164-3d32-432d-9f36-ab2af39527b9\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.346102 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxr82\" (UniqueName: \"kubernetes.io/projected/c736f164-3d32-432d-9f36-ab2af39527b9-kube-api-access-qxr82\") pod \"c736f164-3d32-432d-9f36-ab2af39527b9\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.346317 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c736f164-3d32-432d-9f36-ab2af39527b9-logs\") pod \"c736f164-3d32-432d-9f36-ab2af39527b9\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.346414 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-config-data\") pod \"c736f164-3d32-432d-9f36-ab2af39527b9\" (UID: \"c736f164-3d32-432d-9f36-ab2af39527b9\") " Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.351483 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c736f164-3d32-432d-9f36-ab2af39527b9-logs" (OuterVolumeSpecName: "logs") pod "c736f164-3d32-432d-9f36-ab2af39527b9" (UID: "c736f164-3d32-432d-9f36-ab2af39527b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.365353 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c736f164-3d32-432d-9f36-ab2af39527b9-kube-api-access-qxr82" (OuterVolumeSpecName: "kube-api-access-qxr82") pod "c736f164-3d32-432d-9f36-ab2af39527b9" (UID: "c736f164-3d32-432d-9f36-ab2af39527b9"). InnerVolumeSpecName "kube-api-access-qxr82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.413060 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c736f164-3d32-432d-9f36-ab2af39527b9" (UID: "c736f164-3d32-432d-9f36-ab2af39527b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.449175 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-config-data" (OuterVolumeSpecName: "config-data") pod "c736f164-3d32-432d-9f36-ab2af39527b9" (UID: "c736f164-3d32-432d-9f36-ab2af39527b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.451592 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.451715 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxr82\" (UniqueName: \"kubernetes.io/projected/c736f164-3d32-432d-9f36-ab2af39527b9-kube-api-access-qxr82\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.451810 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c736f164-3d32-432d-9f36-ab2af39527b9-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.451912 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c736f164-3d32-432d-9f36-ab2af39527b9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:42 crc kubenswrapper[4690]: W1203 13:35:42.560749 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98f79021_b882_441a_afb1_52637dd48566.slice/crio-1ef930ba68e9a041c6b12635d299425a92b9c4ab5811bd07fc9f3647e2098552 WatchSource:0}: Error finding container 1ef930ba68e9a041c6b12635d299425a92b9c4ab5811bd07fc9f3647e2098552: Status 404 returned error can't find the container with id 1ef930ba68e9a041c6b12635d299425a92b9c4ab5811bd07fc9f3647e2098552 Dec 03 13:35:42 crc kubenswrapper[4690]: I1203 13:35:42.578300 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnfn6"] Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.188822 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.188821 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c736f164-3d32-432d-9f36-ab2af39527b9","Type":"ContainerDied","Data":"83f2d47ee9689278b1291fcd8a2c2310a0501f7a4c9cd783bfac08745961cfbb"} Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.190425 4690 scope.go:117] "RemoveContainer" containerID="63d55a067ec1a6c03055428043e3f1abbda4194b04511e773d4776774a49d9f4" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.191283 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnfn6" event={"ID":"98f79021-b882-441a-afb1-52637dd48566","Type":"ContainerStarted","Data":"6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe"} Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.191317 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnfn6" event={"ID":"98f79021-b882-441a-afb1-52637dd48566","Type":"ContainerStarted","Data":"1ef930ba68e9a041c6b12635d299425a92b9c4ab5811bd07fc9f3647e2098552"} Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.212309 4690 scope.go:117] "RemoveContainer" containerID="dd74e69f776ba57e0d316bf3a884031e82f948863c6882f0d80a0d5844a2fa6d" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.238862 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.260524 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.273070 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:43 crc kubenswrapper[4690]: E1203 13:35:43.273916 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-log" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.273953 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-log" Dec 03 13:35:43 crc kubenswrapper[4690]: E1203 13:35:43.274058 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-api" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.274075 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-api" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.274438 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-log" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.274504 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" containerName="nova-api-api" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.280247 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.287760 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.297823 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.373331 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ac3ecc8-a288-4892-86f6-63214af5fed0-logs\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.373498 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.373714 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9znr\" (UniqueName: \"kubernetes.io/projected/7ac3ecc8-a288-4892-86f6-63214af5fed0-kube-api-access-k9znr\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.373746 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-config-data\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.475961 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9znr\" (UniqueName: \"kubernetes.io/projected/7ac3ecc8-a288-4892-86f6-63214af5fed0-kube-api-access-k9znr\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.476028 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-config-data\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.476131 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ac3ecc8-a288-4892-86f6-63214af5fed0-logs\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.476224 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.477272 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ac3ecc8-a288-4892-86f6-63214af5fed0-logs\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.484265 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-config-data\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.484265 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.506172 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9znr\" (UniqueName: \"kubernetes.io/projected/7ac3ecc8-a288-4892-86f6-63214af5fed0-kube-api-access-k9znr\") pod \"nova-api-0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " pod="openstack/nova-api-0" Dec 03 13:35:43 crc kubenswrapper[4690]: I1203 13:35:43.606514 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:35:44 crc kubenswrapper[4690]: I1203 13:35:44.085601 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:35:44 crc kubenswrapper[4690]: I1203 13:35:44.202543 4690 generic.go:334] "Generic (PLEG): container finished" podID="98f79021-b882-441a-afb1-52637dd48566" containerID="6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe" exitCode=0 Dec 03 13:35:44 crc kubenswrapper[4690]: I1203 13:35:44.211269 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnfn6" event={"ID":"98f79021-b882-441a-afb1-52637dd48566","Type":"ContainerDied","Data":"6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe"} Dec 03 13:35:44 crc kubenswrapper[4690]: I1203 13:35:44.218157 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ac3ecc8-a288-4892-86f6-63214af5fed0","Type":"ContainerStarted","Data":"72482da2cb4c84c7ee25ba7aabdd7eb2ad58b8c01f871cda70aa76cc1ea32619"} Dec 03 13:35:44 crc kubenswrapper[4690]: I1203 13:35:44.333751 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c736f164-3d32-432d-9f36-ab2af39527b9" path="/var/lib/kubelet/pods/c736f164-3d32-432d-9f36-ab2af39527b9/volumes" Dec 03 13:35:44 crc kubenswrapper[4690]: I1203 13:35:44.486484 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 13:35:45 crc kubenswrapper[4690]: I1203 13:35:45.236948 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ac3ecc8-a288-4892-86f6-63214af5fed0","Type":"ContainerStarted","Data":"18d84ca6973e333c36c4eee80f77b091f60c8f50746155d1777b26780f0d6bbe"} Dec 03 13:35:45 crc kubenswrapper[4690]: I1203 13:35:45.876651 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:35:45 crc kubenswrapper[4690]: I1203 13:35:45.877213 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="7b78850b-78ce-4668-97e8-d4929857b34a" containerName="kube-state-metrics" containerID="cri-o://c4ad4a3b6b0af4f5c90a1321900f91f4b892c5440d86deb04c585b5f395f4036" gracePeriod=30 Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.308250 4690 generic.go:334] "Generic (PLEG): container finished" podID="7b78850b-78ce-4668-97e8-d4929857b34a" containerID="c4ad4a3b6b0af4f5c90a1321900f91f4b892c5440d86deb04c585b5f395f4036" exitCode=2 Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.308743 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7b78850b-78ce-4668-97e8-d4929857b34a","Type":"ContainerDied","Data":"c4ad4a3b6b0af4f5c90a1321900f91f4b892c5440d86deb04c585b5f395f4036"} Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.348388 4690 generic.go:334] "Generic (PLEG): container finished" podID="98f79021-b882-441a-afb1-52637dd48566" containerID="ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6" exitCode=0 Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.364727 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnfn6" event={"ID":"98f79021-b882-441a-afb1-52637dd48566","Type":"ContainerDied","Data":"ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6"} Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.379178 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ac3ecc8-a288-4892-86f6-63214af5fed0","Type":"ContainerStarted","Data":"e01aae61fae6b934f221b25b8b10b939f1b9e20b581859ee43c36cece1f8b1e4"} Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.470932 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.470892775 podStartE2EDuration="3.470892775s" podCreationTimestamp="2025-12-03 13:35:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:35:46.44150345 +0000 UTC m=+1592.422423883" watchObservedRunningTime="2025-12-03 13:35:46.470892775 +0000 UTC m=+1592.451813198" Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.510562 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.557042 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5mlv\" (UniqueName: \"kubernetes.io/projected/7b78850b-78ce-4668-97e8-d4929857b34a-kube-api-access-s5mlv\") pod \"7b78850b-78ce-4668-97e8-d4929857b34a\" (UID: \"7b78850b-78ce-4668-97e8-d4929857b34a\") " Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.579823 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b78850b-78ce-4668-97e8-d4929857b34a-kube-api-access-s5mlv" (OuterVolumeSpecName: "kube-api-access-s5mlv") pod "7b78850b-78ce-4668-97e8-d4929857b34a" (UID: "7b78850b-78ce-4668-97e8-d4929857b34a"). InnerVolumeSpecName "kube-api-access-s5mlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:46 crc kubenswrapper[4690]: I1203 13:35:46.663341 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5mlv\" (UniqueName: \"kubernetes.io/projected/7b78850b-78ce-4668-97e8-d4929857b34a-kube-api-access-s5mlv\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.397725 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7b78850b-78ce-4668-97e8-d4929857b34a","Type":"ContainerDied","Data":"c122fbdbfa57364abcec5cbe90ef1d77c4cb7e66f5a94559a4c3b28c27cc8f50"} Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.398256 4690 scope.go:117] "RemoveContainer" containerID="c4ad4a3b6b0af4f5c90a1321900f91f4b892c5440d86deb04c585b5f395f4036" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.397942 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.409660 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnfn6" event={"ID":"98f79021-b882-441a-afb1-52637dd48566","Type":"ContainerStarted","Data":"0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974"} Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.447970 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jnfn6" podStartSLOduration=3.854006365 podStartE2EDuration="6.447944777s" podCreationTimestamp="2025-12-03 13:35:41 +0000 UTC" firstStartedPulling="2025-12-03 13:35:44.216785647 +0000 UTC m=+1590.197706070" lastFinishedPulling="2025-12-03 13:35:46.810724049 +0000 UTC m=+1592.791644482" observedRunningTime="2025-12-03 13:35:47.431324109 +0000 UTC m=+1593.412244542" watchObservedRunningTime="2025-12-03 13:35:47.447944777 +0000 UTC m=+1593.428865200" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.480591 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.495384 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.513764 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:35:47 crc kubenswrapper[4690]: E1203 13:35:47.514357 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b78850b-78ce-4668-97e8-d4929857b34a" containerName="kube-state-metrics" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.514386 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b78850b-78ce-4668-97e8-d4929857b34a" containerName="kube-state-metrics" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.514623 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b78850b-78ce-4668-97e8-d4929857b34a" containerName="kube-state-metrics" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.515423 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.524172 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.524386 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.529346 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.585904 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw22k\" (UniqueName: \"kubernetes.io/projected/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-api-access-lw22k\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.586012 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.586116 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.586185 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.688576 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.688679 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw22k\" (UniqueName: \"kubernetes.io/projected/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-api-access-lw22k\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.688741 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.688830 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.695001 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.695593 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.696170 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.714643 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw22k\" (UniqueName: \"kubernetes.io/projected/9869e3c2-67c6-47be-8f00-45999ffcd3e7-kube-api-access-lw22k\") pod \"kube-state-metrics-0\" (UID: \"9869e3c2-67c6-47be-8f00-45999ffcd3e7\") " pod="openstack/kube-state-metrics-0" Dec 03 13:35:47 crc kubenswrapper[4690]: I1203 13:35:47.845490 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.190238 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.190528 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="ceilometer-central-agent" containerID="cri-o://842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397" gracePeriod=30 Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.190601 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="proxy-httpd" containerID="cri-o://62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559" gracePeriod=30 Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.190770 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="ceilometer-notification-agent" containerID="cri-o://e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff" gracePeriod=30 Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.190812 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="sg-core" containerID="cri-o://8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8" gracePeriod=30 Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.335210 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b78850b-78ce-4668-97e8-d4929857b34a" path="/var/lib/kubelet/pods/7b78850b-78ce-4668-97e8-d4929857b34a/volumes" Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.431169 4690 generic.go:334] "Generic (PLEG): container finished" podID="7fac3e34-0060-4311-ad94-762e417c578b" containerID="62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559" exitCode=0 Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.431208 4690 generic.go:334] "Generic (PLEG): container finished" podID="7fac3e34-0060-4311-ad94-762e417c578b" containerID="8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8" exitCode=2 Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.432198 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerDied","Data":"62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559"} Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:48.432232 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerDied","Data":"8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8"} Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:49.446438 4690 generic.go:334] "Generic (PLEG): container finished" podID="7fac3e34-0060-4311-ad94-762e417c578b" containerID="842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397" exitCode=0 Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:49.446535 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerDied","Data":"842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397"} Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:49.455411 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:49.486051 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 13:35:49 crc kubenswrapper[4690]: I1203 13:35:49.522461 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 13:35:50 crc kubenswrapper[4690]: I1203 13:35:50.504951 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9869e3c2-67c6-47be-8f00-45999ffcd3e7","Type":"ContainerStarted","Data":"e6aec617bfc9fa99f5cb7fedf7bcaec49e943b67f246d24551d1c3d3c4892b17"} Dec 03 13:35:50 crc kubenswrapper[4690]: I1203 13:35:50.505670 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9869e3c2-67c6-47be-8f00-45999ffcd3e7","Type":"ContainerStarted","Data":"40c8a20575b6252090207e226d842e69d303d0f0087e58feb25ace298d94a73b"} Dec 03 13:35:50 crc kubenswrapper[4690]: I1203 13:35:50.505719 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 13:35:50 crc kubenswrapper[4690]: I1203 13:35:50.536390 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.067512335 podStartE2EDuration="3.536361432s" podCreationTimestamp="2025-12-03 13:35:47 +0000 UTC" firstStartedPulling="2025-12-03 13:35:49.466214285 +0000 UTC m=+1595.447134798" lastFinishedPulling="2025-12-03 13:35:49.935063462 +0000 UTC m=+1595.915983895" observedRunningTime="2025-12-03 13:35:50.523881353 +0000 UTC m=+1596.504801806" watchObservedRunningTime="2025-12-03 13:35:50.536361432 +0000 UTC m=+1596.517281865" Dec 03 13:35:50 crc kubenswrapper[4690]: I1203 13:35:50.546960 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 13:35:50 crc kubenswrapper[4690]: I1203 13:35:50.551656 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 13:35:51 crc kubenswrapper[4690]: I1203 13:35:51.942537 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:51 crc kubenswrapper[4690]: I1203 13:35:51.943080 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:51 crc kubenswrapper[4690]: I1203 13:35:51.994651 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:52 crc kubenswrapper[4690]: I1203 13:35:52.569546 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:52 crc kubenswrapper[4690]: I1203 13:35:52.626918 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnfn6"] Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.024687 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.133059 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-combined-ca-bundle\") pod \"7fac3e34-0060-4311-ad94-762e417c578b\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.133366 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-run-httpd\") pod \"7fac3e34-0060-4311-ad94-762e417c578b\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.134064 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-log-httpd\") pod \"7fac3e34-0060-4311-ad94-762e417c578b\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.134008 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7fac3e34-0060-4311-ad94-762e417c578b" (UID: "7fac3e34-0060-4311-ad94-762e417c578b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.134436 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29lrc\" (UniqueName: \"kubernetes.io/projected/7fac3e34-0060-4311-ad94-762e417c578b-kube-api-access-29lrc\") pod \"7fac3e34-0060-4311-ad94-762e417c578b\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.134564 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7fac3e34-0060-4311-ad94-762e417c578b" (UID: "7fac3e34-0060-4311-ad94-762e417c578b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.134585 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-config-data\") pod \"7fac3e34-0060-4311-ad94-762e417c578b\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.134706 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-sg-core-conf-yaml\") pod \"7fac3e34-0060-4311-ad94-762e417c578b\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.134931 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-scripts\") pod \"7fac3e34-0060-4311-ad94-762e417c578b\" (UID: \"7fac3e34-0060-4311-ad94-762e417c578b\") " Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.136012 4690 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.136187 4690 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fac3e34-0060-4311-ad94-762e417c578b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.155180 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-scripts" (OuterVolumeSpecName: "scripts") pod "7fac3e34-0060-4311-ad94-762e417c578b" (UID: "7fac3e34-0060-4311-ad94-762e417c578b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.155233 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fac3e34-0060-4311-ad94-762e417c578b-kube-api-access-29lrc" (OuterVolumeSpecName: "kube-api-access-29lrc") pod "7fac3e34-0060-4311-ad94-762e417c578b" (UID: "7fac3e34-0060-4311-ad94-762e417c578b"). InnerVolumeSpecName "kube-api-access-29lrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.170979 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7fac3e34-0060-4311-ad94-762e417c578b" (UID: "7fac3e34-0060-4311-ad94-762e417c578b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.223235 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fac3e34-0060-4311-ad94-762e417c578b" (UID: "7fac3e34-0060-4311-ad94-762e417c578b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.238414 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29lrc\" (UniqueName: \"kubernetes.io/projected/7fac3e34-0060-4311-ad94-762e417c578b-kube-api-access-29lrc\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.238460 4690 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.238474 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.238485 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.248218 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-config-data" (OuterVolumeSpecName: "config-data") pod "7fac3e34-0060-4311-ad94-762e417c578b" (UID: "7fac3e34-0060-4311-ad94-762e417c578b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.341410 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fac3e34-0060-4311-ad94-762e417c578b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.540244 4690 generic.go:334] "Generic (PLEG): container finished" podID="7fac3e34-0060-4311-ad94-762e417c578b" containerID="e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff" exitCode=0 Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.540384 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.540365 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerDied","Data":"e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff"} Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.540907 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fac3e34-0060-4311-ad94-762e417c578b","Type":"ContainerDied","Data":"892b432559168d0b6eaf6bff63e4958b39492ac73b720de3ee48d206976f8e14"} Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.540960 4690 scope.go:117] "RemoveContainer" containerID="62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.562661 4690 scope.go:117] "RemoveContainer" containerID="8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.588913 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.600831 4690 scope.go:117] "RemoveContainer" containerID="e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.607192 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.607287 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.607313 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.623939 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:53 crc kubenswrapper[4690]: E1203 13:35:53.624773 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="sg-core" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.624838 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="sg-core" Dec 03 13:35:53 crc kubenswrapper[4690]: E1203 13:35:53.624914 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="ceilometer-notification-agent" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.624969 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="ceilometer-notification-agent" Dec 03 13:35:53 crc kubenswrapper[4690]: E1203 13:35:53.625076 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="ceilometer-central-agent" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.625127 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="ceilometer-central-agent" Dec 03 13:35:53 crc kubenswrapper[4690]: E1203 13:35:53.625181 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="proxy-httpd" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.625226 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="proxy-httpd" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.625515 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="proxy-httpd" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.625585 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="sg-core" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.625649 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="ceilometer-notification-agent" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.625703 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fac3e34-0060-4311-ad94-762e417c578b" containerName="ceilometer-central-agent" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.628217 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.634351 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.637589 4690 scope.go:117] "RemoveContainer" containerID="842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.638596 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.643099 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.644389 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.678433 4690 scope.go:117] "RemoveContainer" containerID="62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559" Dec 03 13:35:53 crc kubenswrapper[4690]: E1203 13:35:53.679005 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559\": container with ID starting with 62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559 not found: ID does not exist" containerID="62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.679081 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559"} err="failed to get container status \"62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559\": rpc error: code = NotFound desc = could not find container \"62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559\": container with ID starting with 62af9144e12d5a5fc7880ed066e38f52332188d54e2fc279c0335b8efb9cc559 not found: ID does not exist" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.679122 4690 scope.go:117] "RemoveContainer" containerID="8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8" Dec 03 13:35:53 crc kubenswrapper[4690]: E1203 13:35:53.679630 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8\": container with ID starting with 8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8 not found: ID does not exist" containerID="8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.679692 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8"} err="failed to get container status \"8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8\": rpc error: code = NotFound desc = could not find container \"8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8\": container with ID starting with 8b0959f3fe26131c365d06f7fabadfbdc986a0f2a2224a35b689addcd6cf2ea8 not found: ID does not exist" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.679738 4690 scope.go:117] "RemoveContainer" containerID="e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff" Dec 03 13:35:53 crc kubenswrapper[4690]: E1203 13:35:53.680906 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff\": container with ID starting with e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff not found: ID does not exist" containerID="e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.680939 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff"} err="failed to get container status \"e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff\": rpc error: code = NotFound desc = could not find container \"e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff\": container with ID starting with e8455145b14f4d254d4891b21afdcfdbe2a85a669490313f063ecdfa20e3dfff not found: ID does not exist" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.680959 4690 scope.go:117] "RemoveContainer" containerID="842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397" Dec 03 13:35:53 crc kubenswrapper[4690]: E1203 13:35:53.681185 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397\": container with ID starting with 842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397 not found: ID does not exist" containerID="842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.681214 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397"} err="failed to get container status \"842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397\": rpc error: code = NotFound desc = could not find container \"842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397\": container with ID starting with 842fbed4ea8048ea137ebb074a17898857331c09cc3f583877cae7219c110397 not found: ID does not exist" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.761857 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvgp5\" (UniqueName: \"kubernetes.io/projected/3ef93d80-8ad3-44f6-8c73-e6f289b75064-kube-api-access-pvgp5\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.764240 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-log-httpd\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.764468 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.765130 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-config-data\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.765365 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-scripts\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.767352 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.767516 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-run-httpd\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.768020 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.871435 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.872065 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-run-httpd\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.872142 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.872195 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvgp5\" (UniqueName: \"kubernetes.io/projected/3ef93d80-8ad3-44f6-8c73-e6f289b75064-kube-api-access-pvgp5\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.872266 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-log-httpd\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.872317 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.872391 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-config-data\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.872462 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-scripts\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.872625 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-run-httpd\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.873570 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-log-httpd\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.878230 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.879246 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-config-data\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.880565 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.884832 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.887602 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-scripts\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.893465 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvgp5\" (UniqueName: \"kubernetes.io/projected/3ef93d80-8ad3-44f6-8c73-e6f289b75064-kube-api-access-pvgp5\") pod \"ceilometer-0\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " pod="openstack/ceilometer-0" Dec 03 13:35:53 crc kubenswrapper[4690]: I1203 13:35:53.958393 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:35:54 crc kubenswrapper[4690]: I1203 13:35:54.331061 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fac3e34-0060-4311-ad94-762e417c578b" path="/var/lib/kubelet/pods/7fac3e34-0060-4311-ad94-762e417c578b/volumes" Dec 03 13:35:54 crc kubenswrapper[4690]: I1203 13:35:54.466785 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:35:54 crc kubenswrapper[4690]: I1203 13:35:54.556175 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerStarted","Data":"0865b751062c0168053859418fbc4bbd7307cf81f1e7aedfa8f9d422e9df0a65"} Dec 03 13:35:54 crc kubenswrapper[4690]: I1203 13:35:54.556366 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jnfn6" podUID="98f79021-b882-441a-afb1-52637dd48566" containerName="registry-server" containerID="cri-o://0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974" gracePeriod=2 Dec 03 13:35:54 crc kubenswrapper[4690]: I1203 13:35:54.690300 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:35:54 crc kubenswrapper[4690]: I1203 13:35:54.690576 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.164726 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.321382 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-catalog-content\") pod \"98f79021-b882-441a-afb1-52637dd48566\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.321665 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdflw\" (UniqueName: \"kubernetes.io/projected/98f79021-b882-441a-afb1-52637dd48566-kube-api-access-wdflw\") pod \"98f79021-b882-441a-afb1-52637dd48566\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.321854 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-utilities\") pod \"98f79021-b882-441a-afb1-52637dd48566\" (UID: \"98f79021-b882-441a-afb1-52637dd48566\") " Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.323704 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-utilities" (OuterVolumeSpecName: "utilities") pod "98f79021-b882-441a-afb1-52637dd48566" (UID: "98f79021-b882-441a-afb1-52637dd48566"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.330050 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98f79021-b882-441a-afb1-52637dd48566-kube-api-access-wdflw" (OuterVolumeSpecName: "kube-api-access-wdflw") pod "98f79021-b882-441a-afb1-52637dd48566" (UID: "98f79021-b882-441a-afb1-52637dd48566"). InnerVolumeSpecName "kube-api-access-wdflw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.347584 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98f79021-b882-441a-afb1-52637dd48566" (UID: "98f79021-b882-441a-afb1-52637dd48566"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.425490 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdflw\" (UniqueName: \"kubernetes.io/projected/98f79021-b882-441a-afb1-52637dd48566-kube-api-access-wdflw\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.425619 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.425674 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f79021-b882-441a-afb1-52637dd48566-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.568161 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerStarted","Data":"d7f4dbd666a894769c3c67a8001bab80ca2542e7f271ae7f804163382d3e201e"} Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.572185 4690 generic.go:334] "Generic (PLEG): container finished" podID="98f79021-b882-441a-afb1-52637dd48566" containerID="0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974" exitCode=0 Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.572214 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnfn6" event={"ID":"98f79021-b882-441a-afb1-52637dd48566","Type":"ContainerDied","Data":"0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974"} Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.572231 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnfn6" event={"ID":"98f79021-b882-441a-afb1-52637dd48566","Type":"ContainerDied","Data":"1ef930ba68e9a041c6b12635d299425a92b9c4ab5811bd07fc9f3647e2098552"} Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.572250 4690 scope.go:117] "RemoveContainer" containerID="0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.572312 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnfn6" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.605835 4690 scope.go:117] "RemoveContainer" containerID="ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.622329 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnfn6"] Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.632340 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnfn6"] Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.646676 4690 scope.go:117] "RemoveContainer" containerID="6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.666639 4690 scope.go:117] "RemoveContainer" containerID="0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974" Dec 03 13:35:55 crc kubenswrapper[4690]: E1203 13:35:55.667440 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974\": container with ID starting with 0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974 not found: ID does not exist" containerID="0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.667469 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974"} err="failed to get container status \"0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974\": rpc error: code = NotFound desc = could not find container \"0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974\": container with ID starting with 0b9505f845e33d67ee184d0709ca01b27a78181c3fdcc0e09d0a2707f62ed974 not found: ID does not exist" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.667490 4690 scope.go:117] "RemoveContainer" containerID="ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6" Dec 03 13:35:55 crc kubenswrapper[4690]: E1203 13:35:55.667915 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6\": container with ID starting with ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6 not found: ID does not exist" containerID="ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.667965 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6"} err="failed to get container status \"ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6\": rpc error: code = NotFound desc = could not find container \"ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6\": container with ID starting with ac1370b940745485fb65ef3bcb562f105ebc41950dcd59f8fdf200e91588a9a6 not found: ID does not exist" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.668003 4690 scope.go:117] "RemoveContainer" containerID="6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe" Dec 03 13:35:55 crc kubenswrapper[4690]: E1203 13:35:55.668400 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe\": container with ID starting with 6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe not found: ID does not exist" containerID="6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe" Dec 03 13:35:55 crc kubenswrapper[4690]: I1203 13:35:55.668439 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe"} err="failed to get container status \"6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe\": rpc error: code = NotFound desc = could not find container \"6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe\": container with ID starting with 6f73e87b03936275601b02048fdd163d9aa5edfdda210be73109c1cd424a4cfe not found: ID does not exist" Dec 03 13:35:56 crc kubenswrapper[4690]: I1203 13:35:56.332514 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98f79021-b882-441a-afb1-52637dd48566" path="/var/lib/kubelet/pods/98f79021-b882-441a-afb1-52637dd48566/volumes" Dec 03 13:35:56 crc kubenswrapper[4690]: I1203 13:35:56.583150 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerStarted","Data":"1d1a933e98eee6a6bbcd5b67fcbc6fb82e401fd0e82855e507ffaae5c08384e9"} Dec 03 13:35:57 crc kubenswrapper[4690]: I1203 13:35:57.598804 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerStarted","Data":"afbda57f7b124d1cd9f10d83cb09473f9bb19fe8a6f5de30aa894c6c956377b7"} Dec 03 13:35:57 crc kubenswrapper[4690]: I1203 13:35:57.856646 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 13:35:59 crc kubenswrapper[4690]: I1203 13:35:59.625831 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerStarted","Data":"28eceada9444dedc03dfb57f7dd3e34f050db178ff268908748efc34795d9496"} Dec 03 13:35:59 crc kubenswrapper[4690]: I1203 13:35:59.626495 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:35:59 crc kubenswrapper[4690]: I1203 13:35:59.656690 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7585250119999998 podStartE2EDuration="6.656664395s" podCreationTimestamp="2025-12-03 13:35:53 +0000 UTC" firstStartedPulling="2025-12-03 13:35:54.471608875 +0000 UTC m=+1600.452529308" lastFinishedPulling="2025-12-03 13:35:58.369748258 +0000 UTC m=+1604.350668691" observedRunningTime="2025-12-03 13:35:59.651541589 +0000 UTC m=+1605.632462032" watchObservedRunningTime="2025-12-03 13:35:59.656664395 +0000 UTC m=+1605.637584828" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.538542 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.548399 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.647775 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjdbr\" (UniqueName: \"kubernetes.io/projected/24bebacb-2b5e-45a7-8fb2-838d4d40193e-kube-api-access-fjdbr\") pod \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.647890 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/7ce6e589-c9fb-40ec-9338-cfe095c1b158-kube-api-access-89vlq\") pod \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.647947 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-config-data\") pod \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.648115 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24bebacb-2b5e-45a7-8fb2-838d4d40193e-logs\") pod \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.648161 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-combined-ca-bundle\") pod \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.648214 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-combined-ca-bundle\") pod \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\" (UID: \"24bebacb-2b5e-45a7-8fb2-838d4d40193e\") " Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.648245 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-config-data\") pod \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\" (UID: \"7ce6e589-c9fb-40ec-9338-cfe095c1b158\") " Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.649257 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24bebacb-2b5e-45a7-8fb2-838d4d40193e-logs" (OuterVolumeSpecName: "logs") pod "24bebacb-2b5e-45a7-8fb2-838d4d40193e" (UID: "24bebacb-2b5e-45a7-8fb2-838d4d40193e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.650437 4690 generic.go:334] "Generic (PLEG): container finished" podID="7ce6e589-c9fb-40ec-9338-cfe095c1b158" containerID="d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6" exitCode=137 Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.650693 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.650760 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7ce6e589-c9fb-40ec-9338-cfe095c1b158","Type":"ContainerDied","Data":"d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6"} Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.650793 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7ce6e589-c9fb-40ec-9338-cfe095c1b158","Type":"ContainerDied","Data":"5b90349fb98269c97361d16508254becba97c6eebbe2944b9709c0f339dbe43c"} Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.650811 4690 scope.go:117] "RemoveContainer" containerID="d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.655829 4690 generic.go:334] "Generic (PLEG): container finished" podID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerID="c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687" exitCode=137 Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.655896 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24bebacb-2b5e-45a7-8fb2-838d4d40193e","Type":"ContainerDied","Data":"c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687"} Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.655934 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24bebacb-2b5e-45a7-8fb2-838d4d40193e","Type":"ContainerDied","Data":"b3c39ff26b3ee64c59aa4eee3578760b896f001eb98c25f2383ea9d1f3b9f7d4"} Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.656020 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.656187 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24bebacb-2b5e-45a7-8fb2-838d4d40193e-kube-api-access-fjdbr" (OuterVolumeSpecName: "kube-api-access-fjdbr") pod "24bebacb-2b5e-45a7-8fb2-838d4d40193e" (UID: "24bebacb-2b5e-45a7-8fb2-838d4d40193e"). InnerVolumeSpecName "kube-api-access-fjdbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.656254 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce6e589-c9fb-40ec-9338-cfe095c1b158-kube-api-access-89vlq" (OuterVolumeSpecName: "kube-api-access-89vlq") pod "7ce6e589-c9fb-40ec-9338-cfe095c1b158" (UID: "7ce6e589-c9fb-40ec-9338-cfe095c1b158"). InnerVolumeSpecName "kube-api-access-89vlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.675143 4690 scope.go:117] "RemoveContainer" containerID="d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6" Dec 03 13:36:01 crc kubenswrapper[4690]: E1203 13:36:01.675694 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6\": container with ID starting with d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6 not found: ID does not exist" containerID="d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.675731 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6"} err="failed to get container status \"d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6\": rpc error: code = NotFound desc = could not find container \"d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6\": container with ID starting with d90e4618899ce8e2b6df700515f1cc561aec9a60d1b3a2ce4d8b257c739f34d6 not found: ID does not exist" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.675753 4690 scope.go:117] "RemoveContainer" containerID="c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.679054 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-config-data" (OuterVolumeSpecName: "config-data") pod "24bebacb-2b5e-45a7-8fb2-838d4d40193e" (UID: "24bebacb-2b5e-45a7-8fb2-838d4d40193e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.684597 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-config-data" (OuterVolumeSpecName: "config-data") pod "7ce6e589-c9fb-40ec-9338-cfe095c1b158" (UID: "7ce6e589-c9fb-40ec-9338-cfe095c1b158"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.687210 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ce6e589-c9fb-40ec-9338-cfe095c1b158" (UID: "7ce6e589-c9fb-40ec-9338-cfe095c1b158"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.705045 4690 scope.go:117] "RemoveContainer" containerID="8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.705047 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24bebacb-2b5e-45a7-8fb2-838d4d40193e" (UID: "24bebacb-2b5e-45a7-8fb2-838d4d40193e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.728176 4690 scope.go:117] "RemoveContainer" containerID="c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687" Dec 03 13:36:01 crc kubenswrapper[4690]: E1203 13:36:01.729431 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687\": container with ID starting with c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687 not found: ID does not exist" containerID="c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.729512 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687"} err="failed to get container status \"c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687\": rpc error: code = NotFound desc = could not find container \"c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687\": container with ID starting with c5c6d8c0f5cd60c7bb3d7724d9f6b3423f173e2cddcd1dc65fc8da756ce66687 not found: ID does not exist" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.729555 4690 scope.go:117] "RemoveContainer" containerID="8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916" Dec 03 13:36:01 crc kubenswrapper[4690]: E1203 13:36:01.730487 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916\": container with ID starting with 8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916 not found: ID does not exist" containerID="8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.730530 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916"} err="failed to get container status \"8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916\": rpc error: code = NotFound desc = could not find container \"8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916\": container with ID starting with 8f7a4b9ee9396c60b34062d4e4d3981e1fd6ca22dd124d5fa82a7ccbf3139916 not found: ID does not exist" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.751261 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjdbr\" (UniqueName: \"kubernetes.io/projected/24bebacb-2b5e-45a7-8fb2-838d4d40193e-kube-api-access-fjdbr\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.751308 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/7ce6e589-c9fb-40ec-9338-cfe095c1b158-kube-api-access-89vlq\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.751323 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.751339 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24bebacb-2b5e-45a7-8fb2-838d4d40193e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.751351 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.751363 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bebacb-2b5e-45a7-8fb2-838d4d40193e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:01 crc kubenswrapper[4690]: I1203 13:36:01.751375 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6e589-c9fb-40ec-9338-cfe095c1b158-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.015944 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.031881 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.043131 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: E1203 13:36:02.043599 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f79021-b882-441a-afb1-52637dd48566" containerName="extract-content" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.043621 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f79021-b882-441a-afb1-52637dd48566" containerName="extract-content" Dec 03 13:36:02 crc kubenswrapper[4690]: E1203 13:36:02.043639 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerName="nova-metadata-log" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.043650 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerName="nova-metadata-log" Dec 03 13:36:02 crc kubenswrapper[4690]: E1203 13:36:02.043665 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f79021-b882-441a-afb1-52637dd48566" containerName="extract-utilities" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.043674 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f79021-b882-441a-afb1-52637dd48566" containerName="extract-utilities" Dec 03 13:36:02 crc kubenswrapper[4690]: E1203 13:36:02.043714 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerName="nova-metadata-metadata" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.043723 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerName="nova-metadata-metadata" Dec 03 13:36:02 crc kubenswrapper[4690]: E1203 13:36:02.043748 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f79021-b882-441a-afb1-52637dd48566" containerName="registry-server" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.043756 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f79021-b882-441a-afb1-52637dd48566" containerName="registry-server" Dec 03 13:36:02 crc kubenswrapper[4690]: E1203 13:36:02.043769 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce6e589-c9fb-40ec-9338-cfe095c1b158" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.043777 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce6e589-c9fb-40ec-9338-cfe095c1b158" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.044100 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="98f79021-b882-441a-afb1-52637dd48566" containerName="registry-server" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.044123 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerName="nova-metadata-log" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.044137 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce6e589-c9fb-40ec-9338-cfe095c1b158" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.044156 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" containerName="nova-metadata-metadata" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.045112 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.047105 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.047849 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.049541 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.053058 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.075744 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.075876 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.075906 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.075975 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.076038 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzlcs\" (UniqueName: \"kubernetes.io/projected/d9db005b-de33-4de4-b7fb-c60783700345-kube-api-access-dzlcs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.093539 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.116817 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.142111 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.144809 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.149795 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.150151 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.155332 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178466 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-config-data\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178540 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178590 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178638 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178685 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178707 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178748 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51224467-f912-4c8a-be4a-8ffafa6df970-logs\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178774 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178819 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzlcs\" (UniqueName: \"kubernetes.io/projected/d9db005b-de33-4de4-b7fb-c60783700345-kube-api-access-dzlcs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.178846 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btrxj\" (UniqueName: \"kubernetes.io/projected/51224467-f912-4c8a-be4a-8ffafa6df970-kube-api-access-btrxj\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.183659 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.184075 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.184777 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.194659 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9db005b-de33-4de4-b7fb-c60783700345-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.201528 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzlcs\" (UniqueName: \"kubernetes.io/projected/d9db005b-de33-4de4-b7fb-c60783700345-kube-api-access-dzlcs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d9db005b-de33-4de4-b7fb-c60783700345\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.281131 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51224467-f912-4c8a-be4a-8ffafa6df970-logs\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.281197 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btrxj\" (UniqueName: \"kubernetes.io/projected/51224467-f912-4c8a-be4a-8ffafa6df970-kube-api-access-btrxj\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.281398 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-config-data\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.281446 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.281485 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.284613 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51224467-f912-4c8a-be4a-8ffafa6df970-logs\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.287116 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.287242 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.289735 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-config-data\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.300392 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btrxj\" (UniqueName: \"kubernetes.io/projected/51224467-f912-4c8a-be4a-8ffafa6df970-kube-api-access-btrxj\") pod \"nova-metadata-0\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.330056 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24bebacb-2b5e-45a7-8fb2-838d4d40193e" path="/var/lib/kubelet/pods/24bebacb-2b5e-45a7-8fb2-838d4d40193e/volumes" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.331307 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce6e589-c9fb-40ec-9338-cfe095c1b158" path="/var/lib/kubelet/pods/7ce6e589-c9fb-40ec-9338-cfe095c1b158/volumes" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.382860 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.468702 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:36:02 crc kubenswrapper[4690]: I1203 13:36:02.892322 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:36:02 crc kubenswrapper[4690]: W1203 13:36:02.893670 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9db005b_de33_4de4_b7fb_c60783700345.slice/crio-c2bbf4cde71656f69f26d6f11a776f659804e60defce79205deb8aebf8eb4e82 WatchSource:0}: Error finding container c2bbf4cde71656f69f26d6f11a776f659804e60defce79205deb8aebf8eb4e82: Status 404 returned error can't find the container with id c2bbf4cde71656f69f26d6f11a776f659804e60defce79205deb8aebf8eb4e82 Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.026836 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.610781 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.611425 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.611470 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.617461 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.686358 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51224467-f912-4c8a-be4a-8ffafa6df970","Type":"ContainerStarted","Data":"faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2"} Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.686401 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51224467-f912-4c8a-be4a-8ffafa6df970","Type":"ContainerStarted","Data":"6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539"} Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.686413 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51224467-f912-4c8a-be4a-8ffafa6df970","Type":"ContainerStarted","Data":"d8fe925fd15dcfd6912ce3f96bf91c7eb04192ae422b19b76196ff230a52ca9f"} Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.691769 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d9db005b-de33-4de4-b7fb-c60783700345","Type":"ContainerStarted","Data":"b340fe3ee4b8d887f81ca6de57c0e8269f7461e5f2433519d00dc4d8a8d970e0"} Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.691814 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.691828 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d9db005b-de33-4de4-b7fb-c60783700345","Type":"ContainerStarted","Data":"c2bbf4cde71656f69f26d6f11a776f659804e60defce79205deb8aebf8eb4e82"} Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.695601 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.722828 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.7227995200000001 podStartE2EDuration="1.72279952s" podCreationTimestamp="2025-12-03 13:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:36:03.713768892 +0000 UTC m=+1609.694689335" watchObservedRunningTime="2025-12-03 13:36:03.72279952 +0000 UTC m=+1609.703719973" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.741412 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.74139047 podStartE2EDuration="1.74139047s" podCreationTimestamp="2025-12-03 13:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:36:03.734257102 +0000 UTC m=+1609.715177535" watchObservedRunningTime="2025-12-03 13:36:03.74139047 +0000 UTC m=+1609.722310913" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.925922 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dphf7"] Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.927633 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.939166 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.939378 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.939529 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.939774 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h44kk\" (UniqueName: \"kubernetes.io/projected/0303b53b-3e79-40cf-8769-65addeca1be9-kube-api-access-h44kk\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.939926 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-config\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.940140 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:03 crc kubenswrapper[4690]: I1203 13:36:03.983143 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dphf7"] Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.049600 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.049712 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.049819 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h44kk\" (UniqueName: \"kubernetes.io/projected/0303b53b-3e79-40cf-8769-65addeca1be9-kube-api-access-h44kk\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.049908 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-config\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.049942 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.049965 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.050910 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.051508 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.054273 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-config\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.054314 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.054934 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.084157 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h44kk\" (UniqueName: \"kubernetes.io/projected/0303b53b-3e79-40cf-8769-65addeca1be9-kube-api-access-h44kk\") pod \"dnsmasq-dns-89c5cd4d5-dphf7\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.255214 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:04 crc kubenswrapper[4690]: I1203 13:36:04.927966 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dphf7"] Dec 03 13:36:05 crc kubenswrapper[4690]: I1203 13:36:05.716490 4690 generic.go:334] "Generic (PLEG): container finished" podID="0303b53b-3e79-40cf-8769-65addeca1be9" containerID="8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854" exitCode=0 Dec 03 13:36:05 crc kubenswrapper[4690]: I1203 13:36:05.716586 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" event={"ID":"0303b53b-3e79-40cf-8769-65addeca1be9","Type":"ContainerDied","Data":"8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854"} Dec 03 13:36:05 crc kubenswrapper[4690]: I1203 13:36:05.717265 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" event={"ID":"0303b53b-3e79-40cf-8769-65addeca1be9","Type":"ContainerStarted","Data":"7f9c48faff043602d087a1d223c2d5f0f718d37c75629ada4ed6bfeccea1a9e6"} Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.427254 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.428072 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="ceilometer-central-agent" containerID="cri-o://d7f4dbd666a894769c3c67a8001bab80ca2542e7f271ae7f804163382d3e201e" gracePeriod=30 Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.428195 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="proxy-httpd" containerID="cri-o://28eceada9444dedc03dfb57f7dd3e34f050db178ff268908748efc34795d9496" gracePeriod=30 Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.428173 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="ceilometer-notification-agent" containerID="cri-o://1d1a933e98eee6a6bbcd5b67fcbc6fb82e401fd0e82855e507ffaae5c08384e9" gracePeriod=30 Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.428159 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="sg-core" containerID="cri-o://afbda57f7b124d1cd9f10d83cb09473f9bb19fe8a6f5de30aa894c6c956377b7" gracePeriod=30 Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.732383 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" event={"ID":"0303b53b-3e79-40cf-8769-65addeca1be9","Type":"ContainerStarted","Data":"7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd"} Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.732517 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.737174 4690 generic.go:334] "Generic (PLEG): container finished" podID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerID="28eceada9444dedc03dfb57f7dd3e34f050db178ff268908748efc34795d9496" exitCode=0 Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.737213 4690 generic.go:334] "Generic (PLEG): container finished" podID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerID="afbda57f7b124d1cd9f10d83cb09473f9bb19fe8a6f5de30aa894c6c956377b7" exitCode=2 Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.737262 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerDied","Data":"28eceada9444dedc03dfb57f7dd3e34f050db178ff268908748efc34795d9496"} Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.737343 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerDied","Data":"afbda57f7b124d1cd9f10d83cb09473f9bb19fe8a6f5de30aa894c6c956377b7"} Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.778663 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" podStartSLOduration=3.778641416 podStartE2EDuration="3.778641416s" podCreationTimestamp="2025-12-03 13:36:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:36:06.76477768 +0000 UTC m=+1612.745698113" watchObservedRunningTime="2025-12-03 13:36:06.778641416 +0000 UTC m=+1612.759561849" Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.846891 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.847274 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-log" containerID="cri-o://18d84ca6973e333c36c4eee80f77b091f60c8f50746155d1777b26780f0d6bbe" gracePeriod=30 Dec 03 13:36:06 crc kubenswrapper[4690]: I1203 13:36:06.847636 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-api" containerID="cri-o://e01aae61fae6b934f221b25b8b10b939f1b9e20b581859ee43c36cece1f8b1e4" gracePeriod=30 Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.384651 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.469556 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.469663 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.765811 4690 generic.go:334] "Generic (PLEG): container finished" podID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerID="1d1a933e98eee6a6bbcd5b67fcbc6fb82e401fd0e82855e507ffaae5c08384e9" exitCode=0 Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.766572 4690 generic.go:334] "Generic (PLEG): container finished" podID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerID="d7f4dbd666a894769c3c67a8001bab80ca2542e7f271ae7f804163382d3e201e" exitCode=0 Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.766025 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerDied","Data":"1d1a933e98eee6a6bbcd5b67fcbc6fb82e401fd0e82855e507ffaae5c08384e9"} Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.766821 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerDied","Data":"d7f4dbd666a894769c3c67a8001bab80ca2542e7f271ae7f804163382d3e201e"} Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.778538 4690 generic.go:334] "Generic (PLEG): container finished" podID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerID="18d84ca6973e333c36c4eee80f77b091f60c8f50746155d1777b26780f0d6bbe" exitCode=143 Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.779282 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ac3ecc8-a288-4892-86f6-63214af5fed0","Type":"ContainerDied","Data":"18d84ca6973e333c36c4eee80f77b091f60c8f50746155d1777b26780f0d6bbe"} Dec 03 13:36:07 crc kubenswrapper[4690]: I1203 13:36:07.975204 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.056903 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-scripts\") pod \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.057017 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-ceilometer-tls-certs\") pod \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.057061 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-combined-ca-bundle\") pod \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.057096 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-config-data\") pod \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.057153 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvgp5\" (UniqueName: \"kubernetes.io/projected/3ef93d80-8ad3-44f6-8c73-e6f289b75064-kube-api-access-pvgp5\") pod \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.057196 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-log-httpd\") pod \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.057361 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-sg-core-conf-yaml\") pod \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.057395 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-run-httpd\") pod \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\" (UID: \"3ef93d80-8ad3-44f6-8c73-e6f289b75064\") " Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.058714 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3ef93d80-8ad3-44f6-8c73-e6f289b75064" (UID: "3ef93d80-8ad3-44f6-8c73-e6f289b75064"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.058984 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3ef93d80-8ad3-44f6-8c73-e6f289b75064" (UID: "3ef93d80-8ad3-44f6-8c73-e6f289b75064"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.067580 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-scripts" (OuterVolumeSpecName: "scripts") pod "3ef93d80-8ad3-44f6-8c73-e6f289b75064" (UID: "3ef93d80-8ad3-44f6-8c73-e6f289b75064"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.080024 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ef93d80-8ad3-44f6-8c73-e6f289b75064-kube-api-access-pvgp5" (OuterVolumeSpecName: "kube-api-access-pvgp5") pod "3ef93d80-8ad3-44f6-8c73-e6f289b75064" (UID: "3ef93d80-8ad3-44f6-8c73-e6f289b75064"). InnerVolumeSpecName "kube-api-access-pvgp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.109014 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3ef93d80-8ad3-44f6-8c73-e6f289b75064" (UID: "3ef93d80-8ad3-44f6-8c73-e6f289b75064"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.148806 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3ef93d80-8ad3-44f6-8c73-e6f289b75064" (UID: "3ef93d80-8ad3-44f6-8c73-e6f289b75064"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.160588 4690 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.160632 4690 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.160646 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.160661 4690 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.160672 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvgp5\" (UniqueName: \"kubernetes.io/projected/3ef93d80-8ad3-44f6-8c73-e6f289b75064-kube-api-access-pvgp5\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.160684 4690 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93d80-8ad3-44f6-8c73-e6f289b75064-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.179764 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ef93d80-8ad3-44f6-8c73-e6f289b75064" (UID: "3ef93d80-8ad3-44f6-8c73-e6f289b75064"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.207890 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-config-data" (OuterVolumeSpecName: "config-data") pod "3ef93d80-8ad3-44f6-8c73-e6f289b75064" (UID: "3ef93d80-8ad3-44f6-8c73-e6f289b75064"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.262576 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.262615 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93d80-8ad3-44f6-8c73-e6f289b75064-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.797819 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93d80-8ad3-44f6-8c73-e6f289b75064","Type":"ContainerDied","Data":"0865b751062c0168053859418fbc4bbd7307cf81f1e7aedfa8f9d422e9df0a65"} Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.797934 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.798478 4690 scope.go:117] "RemoveContainer" containerID="28eceada9444dedc03dfb57f7dd3e34f050db178ff268908748efc34795d9496" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.842684 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.843835 4690 scope.go:117] "RemoveContainer" containerID="afbda57f7b124d1cd9f10d83cb09473f9bb19fe8a6f5de30aa894c6c956377b7" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.863005 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.875928 4690 scope.go:117] "RemoveContainer" containerID="1d1a933e98eee6a6bbcd5b67fcbc6fb82e401fd0e82855e507ffaae5c08384e9" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.878389 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:36:08 crc kubenswrapper[4690]: E1203 13:36:08.879269 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="proxy-httpd" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.879302 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="proxy-httpd" Dec 03 13:36:08 crc kubenswrapper[4690]: E1203 13:36:08.879333 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="sg-core" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.879343 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="sg-core" Dec 03 13:36:08 crc kubenswrapper[4690]: E1203 13:36:08.879365 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="ceilometer-notification-agent" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.879375 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="ceilometer-notification-agent" Dec 03 13:36:08 crc kubenswrapper[4690]: E1203 13:36:08.879411 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="ceilometer-central-agent" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.879421 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="ceilometer-central-agent" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.879728 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="ceilometer-notification-agent" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.879754 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="proxy-httpd" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.879767 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="ceilometer-central-agent" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.879785 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" containerName="sg-core" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.883126 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.890377 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.893526 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.893978 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.900909 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.931311 4690 scope.go:117] "RemoveContainer" containerID="d7f4dbd666a894769c3c67a8001bab80ca2542e7f271ae7f804163382d3e201e" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.978906 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.978989 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6330b5b-9599-49b6-89fa-b890374a7830-run-httpd\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.979029 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6jr\" (UniqueName: \"kubernetes.io/projected/d6330b5b-9599-49b6-89fa-b890374a7830-kube-api-access-7c6jr\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.979136 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6330b5b-9599-49b6-89fa-b890374a7830-log-httpd\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.979207 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.979240 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-scripts\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.979266 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:08 crc kubenswrapper[4690]: I1203 13:36:08.979430 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-config-data\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.081662 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6330b5b-9599-49b6-89fa-b890374a7830-run-httpd\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.081743 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6jr\" (UniqueName: \"kubernetes.io/projected/d6330b5b-9599-49b6-89fa-b890374a7830-kube-api-access-7c6jr\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.081862 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6330b5b-9599-49b6-89fa-b890374a7830-log-httpd\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.081953 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.081985 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-scripts\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.082006 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.082111 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-config-data\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.082265 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.082350 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6330b5b-9599-49b6-89fa-b890374a7830-run-httpd\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.083171 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6330b5b-9599-49b6-89fa-b890374a7830-log-httpd\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.089936 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-scripts\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.096139 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.096217 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.097243 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-config-data\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.105419 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6330b5b-9599-49b6-89fa-b890374a7830-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.106477 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6jr\" (UniqueName: \"kubernetes.io/projected/d6330b5b-9599-49b6-89fa-b890374a7830-kube-api-access-7c6jr\") pod \"ceilometer-0\" (UID: \"d6330b5b-9599-49b6-89fa-b890374a7830\") " pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.223183 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:36:09 crc kubenswrapper[4690]: W1203 13:36:09.715507 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6330b5b_9599_49b6_89fa_b890374a7830.slice/crio-120194d07f585effbf64b8127cc3cd1193490c69527861a794b7afb7e0971ad5 WatchSource:0}: Error finding container 120194d07f585effbf64b8127cc3cd1193490c69527861a794b7afb7e0971ad5: Status 404 returned error can't find the container with id 120194d07f585effbf64b8127cc3cd1193490c69527861a794b7afb7e0971ad5 Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.724823 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:36:09 crc kubenswrapper[4690]: I1203 13:36:09.812115 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6330b5b-9599-49b6-89fa-b890374a7830","Type":"ContainerStarted","Data":"120194d07f585effbf64b8127cc3cd1193490c69527861a794b7afb7e0971ad5"} Dec 03 13:36:10 crc kubenswrapper[4690]: I1203 13:36:10.328334 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ef93d80-8ad3-44f6-8c73-e6f289b75064" path="/var/lib/kubelet/pods/3ef93d80-8ad3-44f6-8c73-e6f289b75064/volumes" Dec 03 13:36:10 crc kubenswrapper[4690]: I1203 13:36:10.830518 4690 generic.go:334] "Generic (PLEG): container finished" podID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerID="e01aae61fae6b934f221b25b8b10b939f1b9e20b581859ee43c36cece1f8b1e4" exitCode=0 Dec 03 13:36:10 crc kubenswrapper[4690]: I1203 13:36:10.830925 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ac3ecc8-a288-4892-86f6-63214af5fed0","Type":"ContainerDied","Data":"e01aae61fae6b934f221b25b8b10b939f1b9e20b581859ee43c36cece1f8b1e4"} Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.102187 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.236795 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ac3ecc8-a288-4892-86f6-63214af5fed0-logs\") pod \"7ac3ecc8-a288-4892-86f6-63214af5fed0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.236856 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9znr\" (UniqueName: \"kubernetes.io/projected/7ac3ecc8-a288-4892-86f6-63214af5fed0-kube-api-access-k9znr\") pod \"7ac3ecc8-a288-4892-86f6-63214af5fed0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.237097 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-config-data\") pod \"7ac3ecc8-a288-4892-86f6-63214af5fed0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.237134 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-combined-ca-bundle\") pod \"7ac3ecc8-a288-4892-86f6-63214af5fed0\" (UID: \"7ac3ecc8-a288-4892-86f6-63214af5fed0\") " Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.237496 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ac3ecc8-a288-4892-86f6-63214af5fed0-logs" (OuterVolumeSpecName: "logs") pod "7ac3ecc8-a288-4892-86f6-63214af5fed0" (UID: "7ac3ecc8-a288-4892-86f6-63214af5fed0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.237956 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ac3ecc8-a288-4892-86f6-63214af5fed0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.243100 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac3ecc8-a288-4892-86f6-63214af5fed0-kube-api-access-k9znr" (OuterVolumeSpecName: "kube-api-access-k9znr") pod "7ac3ecc8-a288-4892-86f6-63214af5fed0" (UID: "7ac3ecc8-a288-4892-86f6-63214af5fed0"). InnerVolumeSpecName "kube-api-access-k9znr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.269180 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ac3ecc8-a288-4892-86f6-63214af5fed0" (UID: "7ac3ecc8-a288-4892-86f6-63214af5fed0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.278363 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-config-data" (OuterVolumeSpecName: "config-data") pod "7ac3ecc8-a288-4892-86f6-63214af5fed0" (UID: "7ac3ecc8-a288-4892-86f6-63214af5fed0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.340198 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9znr\" (UniqueName: \"kubernetes.io/projected/7ac3ecc8-a288-4892-86f6-63214af5fed0-kube-api-access-k9znr\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.340228 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.340238 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac3ecc8-a288-4892-86f6-63214af5fed0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.845327 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6330b5b-9599-49b6-89fa-b890374a7830","Type":"ContainerStarted","Data":"cfafd5933f49139358fcf13c9d591bbdc63640cd90b3e53fda2d7b68852734da"} Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.848651 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ac3ecc8-a288-4892-86f6-63214af5fed0","Type":"ContainerDied","Data":"72482da2cb4c84c7ee25ba7aabdd7eb2ad58b8c01f871cda70aa76cc1ea32619"} Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.848718 4690 scope.go:117] "RemoveContainer" containerID="e01aae61fae6b934f221b25b8b10b939f1b9e20b581859ee43c36cece1f8b1e4" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.848755 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.876127 4690 scope.go:117] "RemoveContainer" containerID="18d84ca6973e333c36c4eee80f77b091f60c8f50746155d1777b26780f0d6bbe" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.889922 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.902498 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.926503 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:11 crc kubenswrapper[4690]: E1203 13:36:11.927516 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-api" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.927537 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-api" Dec 03 13:36:11 crc kubenswrapper[4690]: E1203 13:36:11.927563 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-log" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.927570 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-log" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.927776 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-log" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.927802 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" containerName="nova-api-api" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.928925 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.931508 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.931849 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.932188 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 13:36:11 crc kubenswrapper[4690]: I1203 13:36:11.939128 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.054616 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-config-data\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.054685 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-logs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.054708 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.054883 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.054956 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjpf8\" (UniqueName: \"kubernetes.io/projected/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-kube-api-access-gjpf8\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.055052 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.157052 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.158116 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-config-data\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.158178 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.158204 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-logs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.158371 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.158453 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjpf8\" (UniqueName: \"kubernetes.io/projected/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-kube-api-access-gjpf8\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.158785 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-logs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.162084 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.162088 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.163116 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-config-data\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.163503 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.183939 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjpf8\" (UniqueName: \"kubernetes.io/projected/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-kube-api-access-gjpf8\") pod \"nova-api-0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.251430 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.326389 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac3ecc8-a288-4892-86f6-63214af5fed0" path="/var/lib/kubelet/pods/7ac3ecc8-a288-4892-86f6-63214af5fed0/volumes" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.385226 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.459019 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.469661 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.469714 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.862422 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6330b5b-9599-49b6-89fa-b890374a7830","Type":"ContainerStarted","Data":"1e66c4aa74a9d410591048dbc6fa05ccf0a08123bdb29bcb273167460a5f4082"} Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.882553 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:36:12 crc kubenswrapper[4690]: W1203 13:36:12.921010 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf541fc7d_74a4_44d2_afc7_0e76ec9ce2d0.slice/crio-71e4f329c0edc1e8a69811583518a94773c7cdea65015178537c4e41b0ce88fe WatchSource:0}: Error finding container 71e4f329c0edc1e8a69811583518a94773c7cdea65015178537c4e41b0ce88fe: Status 404 returned error can't find the container with id 71e4f329c0edc1e8a69811583518a94773c7cdea65015178537c4e41b0ce88fe Dec 03 13:36:12 crc kubenswrapper[4690]: I1203 13:36:12.934946 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.102294 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-q95w4"] Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.103783 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.106464 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.106675 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.114112 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-q95w4"] Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.180389 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlvpx\" (UniqueName: \"kubernetes.io/projected/06684437-623f-4562-ac9e-55d65a10b540-kube-api-access-zlvpx\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.180468 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-scripts\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.180488 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-config-data\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.180666 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.282415 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.283196 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlvpx\" (UniqueName: \"kubernetes.io/projected/06684437-623f-4562-ac9e-55d65a10b540-kube-api-access-zlvpx\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.283392 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-scripts\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.283492 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-config-data\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.289522 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.290554 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-scripts\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.297795 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-config-data\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.300397 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlvpx\" (UniqueName: \"kubernetes.io/projected/06684437-623f-4562-ac9e-55d65a10b540-kube-api-access-zlvpx\") pod \"nova-cell1-cell-mapping-q95w4\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.427955 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.483563 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.483749 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.881786 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0","Type":"ContainerStarted","Data":"182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2"} Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.882216 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0","Type":"ContainerStarted","Data":"934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef"} Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.882228 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0","Type":"ContainerStarted","Data":"71e4f329c0edc1e8a69811583518a94773c7cdea65015178537c4e41b0ce88fe"} Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.887113 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6330b5b-9599-49b6-89fa-b890374a7830","Type":"ContainerStarted","Data":"00d44132ea7d76bd35aa883d1c55c42318924dac3ec9f64fa000beead45e5571"} Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.912370 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9123445180000003 podStartE2EDuration="2.912344518s" podCreationTimestamp="2025-12-03 13:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:36:13.906928465 +0000 UTC m=+1619.887848898" watchObservedRunningTime="2025-12-03 13:36:13.912344518 +0000 UTC m=+1619.893264951" Dec 03 13:36:13 crc kubenswrapper[4690]: I1203 13:36:13.938061 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-q95w4"] Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.261116 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.369589 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x5ws6"] Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.370582 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" podUID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerName="dnsmasq-dns" containerID="cri-o://d57dfb990b6a701faa8a781c512958741ff80692d4ce3558add0eaff454a2a27" gracePeriod=10 Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.620442 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" podUID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.203:5353: connect: connection refused" Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.953773 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q95w4" event={"ID":"06684437-623f-4562-ac9e-55d65a10b540","Type":"ContainerStarted","Data":"ef19ceea66a66fa06d0b1457c5a9722986dd5cc078ea0e9177614433aeaf2633"} Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.954337 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q95w4" event={"ID":"06684437-623f-4562-ac9e-55d65a10b540","Type":"ContainerStarted","Data":"336b4114411b99f9e63b27cf6986debc5532b708b5e440e28f4e05a8942aaf5a"} Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.985794 4690 generic.go:334] "Generic (PLEG): container finished" podID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerID="d57dfb990b6a701faa8a781c512958741ff80692d4ce3558add0eaff454a2a27" exitCode=0 Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.987178 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" event={"ID":"7290142a-86a6-45d7-8f74-98fb5bf783ed","Type":"ContainerDied","Data":"d57dfb990b6a701faa8a781c512958741ff80692d4ce3558add0eaff454a2a27"} Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.987238 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" event={"ID":"7290142a-86a6-45d7-8f74-98fb5bf783ed","Type":"ContainerDied","Data":"1ce909fe3b2ad7e748b5eaa9bf20d1f0bcc641dcadbaa832c56a7e4682e5f009"} Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.987253 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ce909fe3b2ad7e748b5eaa9bf20d1f0bcc641dcadbaa832c56a7e4682e5f009" Dec 03 13:36:14 crc kubenswrapper[4690]: I1203 13:36:14.995670 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-q95w4" podStartSLOduration=1.995639661 podStartE2EDuration="1.995639661s" podCreationTimestamp="2025-12-03 13:36:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:36:14.975098975 +0000 UTC m=+1620.956019418" watchObservedRunningTime="2025-12-03 13:36:14.995639661 +0000 UTC m=+1620.976560094" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.096793 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.151973 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4xnz\" (UniqueName: \"kubernetes.io/projected/7290142a-86a6-45d7-8f74-98fb5bf783ed-kube-api-access-q4xnz\") pod \"7290142a-86a6-45d7-8f74-98fb5bf783ed\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.152215 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-nb\") pod \"7290142a-86a6-45d7-8f74-98fb5bf783ed\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.152281 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-config\") pod \"7290142a-86a6-45d7-8f74-98fb5bf783ed\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.152314 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-sb\") pod \"7290142a-86a6-45d7-8f74-98fb5bf783ed\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.152409 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-swift-storage-0\") pod \"7290142a-86a6-45d7-8f74-98fb5bf783ed\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.157845 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7290142a-86a6-45d7-8f74-98fb5bf783ed-kube-api-access-q4xnz" (OuterVolumeSpecName: "kube-api-access-q4xnz") pod "7290142a-86a6-45d7-8f74-98fb5bf783ed" (UID: "7290142a-86a6-45d7-8f74-98fb5bf783ed"). InnerVolumeSpecName "kube-api-access-q4xnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.255469 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-svc\") pod \"7290142a-86a6-45d7-8f74-98fb5bf783ed\" (UID: \"7290142a-86a6-45d7-8f74-98fb5bf783ed\") " Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.264993 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4xnz\" (UniqueName: \"kubernetes.io/projected/7290142a-86a6-45d7-8f74-98fb5bf783ed-kube-api-access-q4xnz\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.294640 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7290142a-86a6-45d7-8f74-98fb5bf783ed" (UID: "7290142a-86a6-45d7-8f74-98fb5bf783ed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.295646 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7290142a-86a6-45d7-8f74-98fb5bf783ed" (UID: "7290142a-86a6-45d7-8f74-98fb5bf783ed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.314973 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-config" (OuterVolumeSpecName: "config") pod "7290142a-86a6-45d7-8f74-98fb5bf783ed" (UID: "7290142a-86a6-45d7-8f74-98fb5bf783ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.328455 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7290142a-86a6-45d7-8f74-98fb5bf783ed" (UID: "7290142a-86a6-45d7-8f74-98fb5bf783ed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.344395 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7290142a-86a6-45d7-8f74-98fb5bf783ed" (UID: "7290142a-86a6-45d7-8f74-98fb5bf783ed"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.367614 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.367659 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.367670 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.367679 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:15 crc kubenswrapper[4690]: I1203 13:36:15.367690 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7290142a-86a6-45d7-8f74-98fb5bf783ed-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:16 crc kubenswrapper[4690]: I1203 13:36:16.002997 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6330b5b-9599-49b6-89fa-b890374a7830","Type":"ContainerStarted","Data":"c6fd231a8a39841f06488d9688e1a8be9aff0153739774cc952ba90d38a9c335"} Dec 03 13:36:16 crc kubenswrapper[4690]: I1203 13:36:16.003055 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x5ws6" Dec 03 13:36:16 crc kubenswrapper[4690]: I1203 13:36:16.098359 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.908697073 podStartE2EDuration="8.098340501s" podCreationTimestamp="2025-12-03 13:36:08 +0000 UTC" firstStartedPulling="2025-12-03 13:36:09.718283896 +0000 UTC m=+1615.699204329" lastFinishedPulling="2025-12-03 13:36:14.907927324 +0000 UTC m=+1620.888847757" observedRunningTime="2025-12-03 13:36:16.076278866 +0000 UTC m=+1622.057199299" watchObservedRunningTime="2025-12-03 13:36:16.098340501 +0000 UTC m=+1622.079260934" Dec 03 13:36:16 crc kubenswrapper[4690]: I1203 13:36:16.117575 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x5ws6"] Dec 03 13:36:16 crc kubenswrapper[4690]: I1203 13:36:16.133534 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x5ws6"] Dec 03 13:36:16 crc kubenswrapper[4690]: I1203 13:36:16.332393 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7290142a-86a6-45d7-8f74-98fb5bf783ed" path="/var/lib/kubelet/pods/7290142a-86a6-45d7-8f74-98fb5bf783ed/volumes" Dec 03 13:36:16 crc kubenswrapper[4690]: I1203 13:36:16.824213 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:36:16 crc kubenswrapper[4690]: I1203 13:36:16.824298 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:36:17 crc kubenswrapper[4690]: I1203 13:36:17.013317 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:36:20 crc kubenswrapper[4690]: I1203 13:36:20.044544 4690 generic.go:334] "Generic (PLEG): container finished" podID="06684437-623f-4562-ac9e-55d65a10b540" containerID="ef19ceea66a66fa06d0b1457c5a9722986dd5cc078ea0e9177614433aeaf2633" exitCode=0 Dec 03 13:36:20 crc kubenswrapper[4690]: I1203 13:36:20.044623 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q95w4" event={"ID":"06684437-623f-4562-ac9e-55d65a10b540","Type":"ContainerDied","Data":"ef19ceea66a66fa06d0b1457c5a9722986dd5cc078ea0e9177614433aeaf2633"} Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.506441 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.626210 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlvpx\" (UniqueName: \"kubernetes.io/projected/06684437-623f-4562-ac9e-55d65a10b540-kube-api-access-zlvpx\") pod \"06684437-623f-4562-ac9e-55d65a10b540\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.626301 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-combined-ca-bundle\") pod \"06684437-623f-4562-ac9e-55d65a10b540\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.626372 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-scripts\") pod \"06684437-623f-4562-ac9e-55d65a10b540\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.627384 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-config-data\") pod \"06684437-623f-4562-ac9e-55d65a10b540\" (UID: \"06684437-623f-4562-ac9e-55d65a10b540\") " Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.633788 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-scripts" (OuterVolumeSpecName: "scripts") pod "06684437-623f-4562-ac9e-55d65a10b540" (UID: "06684437-623f-4562-ac9e-55d65a10b540"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.634309 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06684437-623f-4562-ac9e-55d65a10b540-kube-api-access-zlvpx" (OuterVolumeSpecName: "kube-api-access-zlvpx") pod "06684437-623f-4562-ac9e-55d65a10b540" (UID: "06684437-623f-4562-ac9e-55d65a10b540"). InnerVolumeSpecName "kube-api-access-zlvpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.664075 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06684437-623f-4562-ac9e-55d65a10b540" (UID: "06684437-623f-4562-ac9e-55d65a10b540"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.668791 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-config-data" (OuterVolumeSpecName: "config-data") pod "06684437-623f-4562-ac9e-55d65a10b540" (UID: "06684437-623f-4562-ac9e-55d65a10b540"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.729699 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlvpx\" (UniqueName: \"kubernetes.io/projected/06684437-623f-4562-ac9e-55d65a10b540-kube-api-access-zlvpx\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.729938 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.729998 4690 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:21 crc kubenswrapper[4690]: I1203 13:36:21.730048 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06684437-623f-4562-ac9e-55d65a10b540-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.074304 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q95w4" event={"ID":"06684437-623f-4562-ac9e-55d65a10b540","Type":"ContainerDied","Data":"336b4114411b99f9e63b27cf6986debc5532b708b5e440e28f4e05a8942aaf5a"} Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.074408 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="336b4114411b99f9e63b27cf6986debc5532b708b5e440e28f4e05a8942aaf5a" Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.074504 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q95w4" Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.252053 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.252122 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.353369 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.353713 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="345cf534-66c9-430b-b919-25add004cfdc" containerName="nova-scheduler-scheduler" containerID="cri-o://e2deefc31f7e41d9f1c361ab9de4f03d2af25682181758a2fd11d0ceaac50882" gracePeriod=30 Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.365065 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.425802 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.426119 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-log" containerID="cri-o://6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539" gracePeriod=30 Dec 03 13:36:22 crc kubenswrapper[4690]: I1203 13:36:22.426474 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-metadata" containerID="cri-o://faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2" gracePeriod=30 Dec 03 13:36:23 crc kubenswrapper[4690]: I1203 13:36:23.088084 4690 generic.go:334] "Generic (PLEG): container finished" podID="51224467-f912-4c8a-be4a-8ffafa6df970" containerID="6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539" exitCode=143 Dec 03 13:36:23 crc kubenswrapper[4690]: I1203 13:36:23.088175 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51224467-f912-4c8a-be4a-8ffafa6df970","Type":"ContainerDied","Data":"6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539"} Dec 03 13:36:23 crc kubenswrapper[4690]: I1203 13:36:23.088828 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-log" containerID="cri-o://934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef" gracePeriod=30 Dec 03 13:36:23 crc kubenswrapper[4690]: I1203 13:36:23.089831 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-api" containerID="cri-o://182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2" gracePeriod=30 Dec 03 13:36:23 crc kubenswrapper[4690]: I1203 13:36:23.098237 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.215:8774/\": EOF" Dec 03 13:36:23 crc kubenswrapper[4690]: I1203 13:36:23.103968 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.215:8774/\": EOF" Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.108719 4690 generic.go:334] "Generic (PLEG): container finished" podID="345cf534-66c9-430b-b919-25add004cfdc" containerID="e2deefc31f7e41d9f1c361ab9de4f03d2af25682181758a2fd11d0ceaac50882" exitCode=0 Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.108793 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"345cf534-66c9-430b-b919-25add004cfdc","Type":"ContainerDied","Data":"e2deefc31f7e41d9f1c361ab9de4f03d2af25682181758a2fd11d0ceaac50882"} Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.113059 4690 generic.go:334] "Generic (PLEG): container finished" podID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerID="934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef" exitCode=143 Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.113099 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0","Type":"ContainerDied","Data":"934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef"} Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.449968 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.604116 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-combined-ca-bundle\") pod \"345cf534-66c9-430b-b919-25add004cfdc\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.604309 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrpgr\" (UniqueName: \"kubernetes.io/projected/345cf534-66c9-430b-b919-25add004cfdc-kube-api-access-wrpgr\") pod \"345cf534-66c9-430b-b919-25add004cfdc\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.604409 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-config-data\") pod \"345cf534-66c9-430b-b919-25add004cfdc\" (UID: \"345cf534-66c9-430b-b919-25add004cfdc\") " Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.611706 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345cf534-66c9-430b-b919-25add004cfdc-kube-api-access-wrpgr" (OuterVolumeSpecName: "kube-api-access-wrpgr") pod "345cf534-66c9-430b-b919-25add004cfdc" (UID: "345cf534-66c9-430b-b919-25add004cfdc"). InnerVolumeSpecName "kube-api-access-wrpgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.650300 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "345cf534-66c9-430b-b919-25add004cfdc" (UID: "345cf534-66c9-430b-b919-25add004cfdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.656778 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-config-data" (OuterVolumeSpecName: "config-data") pod "345cf534-66c9-430b-b919-25add004cfdc" (UID: "345cf534-66c9-430b-b919-25add004cfdc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.706917 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.706950 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrpgr\" (UniqueName: \"kubernetes.io/projected/345cf534-66c9-430b-b919-25add004cfdc-kube-api-access-wrpgr\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:24 crc kubenswrapper[4690]: I1203 13:36:24.706966 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345cf534-66c9-430b-b919-25add004cfdc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.127518 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"345cf534-66c9-430b-b919-25add004cfdc","Type":"ContainerDied","Data":"0ab43219c13031c5ae8795210fa06f45f080ee62db1a6f959e154a6995ab3a2c"} Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.127587 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.127601 4690 scope.go:117] "RemoveContainer" containerID="e2deefc31f7e41d9f1c361ab9de4f03d2af25682181758a2fd11d0ceaac50882" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.187330 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.201437 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.217695 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:36:25 crc kubenswrapper[4690]: E1203 13:36:25.218351 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345cf534-66c9-430b-b919-25add004cfdc" containerName="nova-scheduler-scheduler" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.218380 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="345cf534-66c9-430b-b919-25add004cfdc" containerName="nova-scheduler-scheduler" Dec 03 13:36:25 crc kubenswrapper[4690]: E1203 13:36:25.218396 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerName="dnsmasq-dns" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.218404 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerName="dnsmasq-dns" Dec 03 13:36:25 crc kubenswrapper[4690]: E1203 13:36:25.218424 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06684437-623f-4562-ac9e-55d65a10b540" containerName="nova-manage" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.218435 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="06684437-623f-4562-ac9e-55d65a10b540" containerName="nova-manage" Dec 03 13:36:25 crc kubenswrapper[4690]: E1203 13:36:25.218461 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerName="init" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.218469 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerName="init" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.218735 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="7290142a-86a6-45d7-8f74-98fb5bf783ed" containerName="dnsmasq-dns" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.218754 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="06684437-623f-4562-ac9e-55d65a10b540" containerName="nova-manage" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.218775 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="345cf534-66c9-430b-b919-25add004cfdc" containerName="nova-scheduler-scheduler" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.219810 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.222288 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.231844 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.320071 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzzmc\" (UniqueName: \"kubernetes.io/projected/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-kube-api-access-jzzmc\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.320373 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-config-data\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.320441 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.424191 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.424446 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzzmc\" (UniqueName: \"kubernetes.io/projected/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-kube-api-access-jzzmc\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.424544 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-config-data\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.438738 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.438897 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-config-data\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.443387 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzzmc\" (UniqueName: \"kubernetes.io/projected/cc0fda9d-61ad-4a67-8c22-1da2e41949b0-kube-api-access-jzzmc\") pod \"nova-scheduler-0\" (UID: \"cc0fda9d-61ad-4a67-8c22-1da2e41949b0\") " pod="openstack/nova-scheduler-0" Dec 03 13:36:25 crc kubenswrapper[4690]: I1203 13:36:25.542975 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.053202 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:36:26 crc kubenswrapper[4690]: W1203 13:36:26.061155 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc0fda9d_61ad_4a67_8c22_1da2e41949b0.slice/crio-4eca9278335d5c3e4a4120865b0d1fbf77c06b588ccccf63ea245c0b2086ad9c WatchSource:0}: Error finding container 4eca9278335d5c3e4a4120865b0d1fbf77c06b588ccccf63ea245c0b2086ad9c: Status 404 returned error can't find the container with id 4eca9278335d5c3e4a4120865b0d1fbf77c06b588ccccf63ea245c0b2086ad9c Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.073185 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.142417 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-config-data\") pod \"51224467-f912-4c8a-be4a-8ffafa6df970\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.142514 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-nova-metadata-tls-certs\") pod \"51224467-f912-4c8a-be4a-8ffafa6df970\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.142584 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51224467-f912-4c8a-be4a-8ffafa6df970-logs\") pod \"51224467-f912-4c8a-be4a-8ffafa6df970\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.142719 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-combined-ca-bundle\") pod \"51224467-f912-4c8a-be4a-8ffafa6df970\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.142799 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btrxj\" (UniqueName: \"kubernetes.io/projected/51224467-f912-4c8a-be4a-8ffafa6df970-kube-api-access-btrxj\") pod \"51224467-f912-4c8a-be4a-8ffafa6df970\" (UID: \"51224467-f912-4c8a-be4a-8ffafa6df970\") " Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.143761 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51224467-f912-4c8a-be4a-8ffafa6df970-logs" (OuterVolumeSpecName: "logs") pod "51224467-f912-4c8a-be4a-8ffafa6df970" (UID: "51224467-f912-4c8a-be4a-8ffafa6df970"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.148779 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51224467-f912-4c8a-be4a-8ffafa6df970-kube-api-access-btrxj" (OuterVolumeSpecName: "kube-api-access-btrxj") pod "51224467-f912-4c8a-be4a-8ffafa6df970" (UID: "51224467-f912-4c8a-be4a-8ffafa6df970"). InnerVolumeSpecName "kube-api-access-btrxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.157169 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc0fda9d-61ad-4a67-8c22-1da2e41949b0","Type":"ContainerStarted","Data":"4eca9278335d5c3e4a4120865b0d1fbf77c06b588ccccf63ea245c0b2086ad9c"} Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.160082 4690 generic.go:334] "Generic (PLEG): container finished" podID="51224467-f912-4c8a-be4a-8ffafa6df970" containerID="faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2" exitCode=0 Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.160149 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51224467-f912-4c8a-be4a-8ffafa6df970","Type":"ContainerDied","Data":"faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2"} Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.160179 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51224467-f912-4c8a-be4a-8ffafa6df970","Type":"ContainerDied","Data":"d8fe925fd15dcfd6912ce3f96bf91c7eb04192ae422b19b76196ff230a52ca9f"} Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.160198 4690 scope.go:117] "RemoveContainer" containerID="faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.160349 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.178843 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51224467-f912-4c8a-be4a-8ffafa6df970" (UID: "51224467-f912-4c8a-be4a-8ffafa6df970"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.185984 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-config-data" (OuterVolumeSpecName: "config-data") pod "51224467-f912-4c8a-be4a-8ffafa6df970" (UID: "51224467-f912-4c8a-be4a-8ffafa6df970"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.208774 4690 scope.go:117] "RemoveContainer" containerID="6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.230043 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "51224467-f912-4c8a-be4a-8ffafa6df970" (UID: "51224467-f912-4c8a-be4a-8ffafa6df970"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.230205 4690 scope.go:117] "RemoveContainer" containerID="faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2" Dec 03 13:36:26 crc kubenswrapper[4690]: E1203 13:36:26.233030 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2\": container with ID starting with faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2 not found: ID does not exist" containerID="faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.233069 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2"} err="failed to get container status \"faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2\": rpc error: code = NotFound desc = could not find container \"faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2\": container with ID starting with faae325faed95c2379f04a817adb5f3a71e71fbdfb8f18245be02c42025956a2 not found: ID does not exist" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.233090 4690 scope.go:117] "RemoveContainer" containerID="6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539" Dec 03 13:36:26 crc kubenswrapper[4690]: E1203 13:36:26.233552 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539\": container with ID starting with 6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539 not found: ID does not exist" containerID="6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.233573 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539"} err="failed to get container status \"6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539\": rpc error: code = NotFound desc = could not find container \"6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539\": container with ID starting with 6773b287d6356a91c99ea08de8898c64ab36e8572693273b5cbc3828cf18a539 not found: ID does not exist" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.245743 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.245775 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btrxj\" (UniqueName: \"kubernetes.io/projected/51224467-f912-4c8a-be4a-8ffafa6df970-kube-api-access-btrxj\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.245788 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.245799 4690 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/51224467-f912-4c8a-be4a-8ffafa6df970-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.245810 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51224467-f912-4c8a-be4a-8ffafa6df970-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.327145 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345cf534-66c9-430b-b919-25add004cfdc" path="/var/lib/kubelet/pods/345cf534-66c9-430b-b919-25add004cfdc/volumes" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.529054 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.555118 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.589194 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:26 crc kubenswrapper[4690]: E1203 13:36:26.604371 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-metadata" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.604410 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-metadata" Dec 03 13:36:26 crc kubenswrapper[4690]: E1203 13:36:26.604443 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-log" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.604451 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-log" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.604735 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-log" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.604767 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" containerName="nova-metadata-metadata" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.606497 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.609028 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.609271 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.609436 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.756824 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg4rk\" (UniqueName: \"kubernetes.io/projected/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-kube-api-access-xg4rk\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.756981 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.757014 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-logs\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.757068 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-config-data\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.757180 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.859439 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg4rk\" (UniqueName: \"kubernetes.io/projected/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-kube-api-access-xg4rk\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.859856 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.859896 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-logs\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.859927 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-config-data\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.859972 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.860634 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-logs\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.866017 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.866282 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.875402 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg4rk\" (UniqueName: \"kubernetes.io/projected/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-kube-api-access-xg4rk\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.888762 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f0b2df1-b5ed-498c-b6dd-784a32f97b3f-config-data\") pod \"nova-metadata-0\" (UID: \"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f\") " pod="openstack/nova-metadata-0" Dec 03 13:36:26 crc kubenswrapper[4690]: I1203 13:36:26.934921 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:36:27 crc kubenswrapper[4690]: I1203 13:36:27.190242 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc0fda9d-61ad-4a67-8c22-1da2e41949b0","Type":"ContainerStarted","Data":"c895bcb59cba917679e62a9677c17be6918bda3c6a8f2021118f1086161ad472"} Dec 03 13:36:27 crc kubenswrapper[4690]: I1203 13:36:27.209923 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.209903792 podStartE2EDuration="2.209903792s" podCreationTimestamp="2025-12-03 13:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:36:27.209592224 +0000 UTC m=+1633.190512657" watchObservedRunningTime="2025-12-03 13:36:27.209903792 +0000 UTC m=+1633.190824215" Dec 03 13:36:27 crc kubenswrapper[4690]: I1203 13:36:27.435719 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:36:27 crc kubenswrapper[4690]: W1203 13:36:27.444318 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f0b2df1_b5ed_498c_b6dd_784a32f97b3f.slice/crio-0e425ea8a83ecfb1b5b36493bd9490e395fa7b9626c3a30aefaae92b6aed6898 WatchSource:0}: Error finding container 0e425ea8a83ecfb1b5b36493bd9490e395fa7b9626c3a30aefaae92b6aed6898: Status 404 returned error can't find the container with id 0e425ea8a83ecfb1b5b36493bd9490e395fa7b9626c3a30aefaae92b6aed6898 Dec 03 13:36:28 crc kubenswrapper[4690]: I1203 13:36:28.212128 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f","Type":"ContainerStarted","Data":"26fd591472a6cb4e650264c20e46fb622a530e4249bf11272e2e61f1606b9566"} Dec 03 13:36:28 crc kubenswrapper[4690]: I1203 13:36:28.212548 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f","Type":"ContainerStarted","Data":"a2f3cf4d42097f7d46183b937c17b68ad8651f6da7eccb771d35ee91b59b07f1"} Dec 03 13:36:28 crc kubenswrapper[4690]: I1203 13:36:28.212573 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3f0b2df1-b5ed-498c-b6dd-784a32f97b3f","Type":"ContainerStarted","Data":"0e425ea8a83ecfb1b5b36493bd9490e395fa7b9626c3a30aefaae92b6aed6898"} Dec 03 13:36:28 crc kubenswrapper[4690]: I1203 13:36:28.240799 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.240776521 podStartE2EDuration="2.240776521s" podCreationTimestamp="2025-12-03 13:36:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:36:28.231441738 +0000 UTC m=+1634.212362201" watchObservedRunningTime="2025-12-03 13:36:28.240776521 +0000 UTC m=+1634.221696964" Dec 03 13:36:28 crc kubenswrapper[4690]: I1203 13:36:28.334709 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51224467-f912-4c8a-be4a-8ffafa6df970" path="/var/lib/kubelet/pods/51224467-f912-4c8a-be4a-8ffafa6df970/volumes" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.013517 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.137615 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjpf8\" (UniqueName: \"kubernetes.io/projected/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-kube-api-access-gjpf8\") pod \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.138108 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-logs\") pod \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.138155 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-public-tls-certs\") pod \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.138293 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-combined-ca-bundle\") pod \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.138357 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-config-data\") pod \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.138406 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-internal-tls-certs\") pod \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\" (UID: \"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0\") " Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.138855 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-logs" (OuterVolumeSpecName: "logs") pod "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" (UID: "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.139222 4690 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.151922 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-kube-api-access-gjpf8" (OuterVolumeSpecName: "kube-api-access-gjpf8") pod "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" (UID: "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0"). InnerVolumeSpecName "kube-api-access-gjpf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.169750 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-config-data" (OuterVolumeSpecName: "config-data") pod "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" (UID: "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.176260 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" (UID: "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.202630 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" (UID: "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.205968 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" (UID: "f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.224975 4690 generic.go:334] "Generic (PLEG): container finished" podID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerID="182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2" exitCode=0 Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.225270 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.225271 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0","Type":"ContainerDied","Data":"182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2"} Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.225331 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0","Type":"ContainerDied","Data":"71e4f329c0edc1e8a69811583518a94773c7cdea65015178537c4e41b0ce88fe"} Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.225379 4690 scope.go:117] "RemoveContainer" containerID="182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.240980 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjpf8\" (UniqueName: \"kubernetes.io/projected/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-kube-api-access-gjpf8\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.241004 4690 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.241016 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.241025 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.241035 4690 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.286926 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.294592 4690 scope.go:117] "RemoveContainer" containerID="934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.297658 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.315617 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:29 crc kubenswrapper[4690]: E1203 13:36:29.316124 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-log" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.316143 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-log" Dec 03 13:36:29 crc kubenswrapper[4690]: E1203 13:36:29.316162 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-api" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.316168 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-api" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.316374 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-api" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.316403 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" containerName="nova-api-log" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.317558 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.320328 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.320566 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.321774 4690 scope.go:117] "RemoveContainer" containerID="182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2" Dec 03 13:36:29 crc kubenswrapper[4690]: E1203 13:36:29.322441 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2\": container with ID starting with 182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2 not found: ID does not exist" containerID="182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.322470 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2"} err="failed to get container status \"182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2\": rpc error: code = NotFound desc = could not find container \"182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2\": container with ID starting with 182fd23406378f98f20be9681fae4f33553a9a17ad5b8bd044b6db1b7af7c6a2 not found: ID does not exist" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.322496 4690 scope.go:117] "RemoveContainer" containerID="934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef" Dec 03 13:36:29 crc kubenswrapper[4690]: E1203 13:36:29.323248 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef\": container with ID starting with 934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef not found: ID does not exist" containerID="934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.323266 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef"} err="failed to get container status \"934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef\": rpc error: code = NotFound desc = could not find container \"934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef\": container with ID starting with 934c5bd6c1e75ce0a31f3bfbf239115436aa6c41b65a9249d997128ecf8147ef not found: ID does not exist" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.323383 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.340446 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.445068 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-config-data\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.445162 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnsqq\" (UniqueName: \"kubernetes.io/projected/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-kube-api-access-dnsqq\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.445336 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.445402 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-logs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.445419 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-public-tls-certs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.445436 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.547067 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.547198 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-logs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.547234 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-public-tls-certs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.547268 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.547355 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-config-data\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.547407 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnsqq\" (UniqueName: \"kubernetes.io/projected/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-kube-api-access-dnsqq\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.548246 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-logs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.552257 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.553057 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-config-data\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.553729 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-public-tls-certs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.553938 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.568218 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnsqq\" (UniqueName: \"kubernetes.io/projected/9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278-kube-api-access-dnsqq\") pod \"nova-api-0\" (UID: \"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278\") " pod="openstack/nova-api-0" Dec 03 13:36:29 crc kubenswrapper[4690]: I1203 13:36:29.644698 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:36:30 crc kubenswrapper[4690]: I1203 13:36:30.091347 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:36:30 crc kubenswrapper[4690]: I1203 13:36:30.238206 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278","Type":"ContainerStarted","Data":"f367e1c8cff4824a6747dadfcafbd2799a9f346a3ecef0474d9b2dce1e7e0407"} Dec 03 13:36:30 crc kubenswrapper[4690]: I1203 13:36:30.330952 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0" path="/var/lib/kubelet/pods/f541fc7d-74a4-44d2-afc7-0e76ec9ce2d0/volumes" Dec 03 13:36:30 crc kubenswrapper[4690]: I1203 13:36:30.543973 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 13:36:31 crc kubenswrapper[4690]: I1203 13:36:31.251255 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278","Type":"ContainerStarted","Data":"24d26f61a0b90322f0de8ad4ce3587d9869607f3962d576eef518c67b2da7e48"} Dec 03 13:36:31 crc kubenswrapper[4690]: I1203 13:36:31.251682 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278","Type":"ContainerStarted","Data":"df0187dede4f0b504420f5a3db0a9b5b856d338a3cd7a96859591b280dd2f2a1"} Dec 03 13:36:31 crc kubenswrapper[4690]: I1203 13:36:31.276518 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.276493294 podStartE2EDuration="2.276493294s" podCreationTimestamp="2025-12-03 13:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:36:31.269037429 +0000 UTC m=+1637.249957872" watchObservedRunningTime="2025-12-03 13:36:31.276493294 +0000 UTC m=+1637.257413727" Dec 03 13:36:31 crc kubenswrapper[4690]: I1203 13:36:31.935348 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:36:31 crc kubenswrapper[4690]: I1203 13:36:31.935398 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:36:35 crc kubenswrapper[4690]: I1203 13:36:35.544046 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 13:36:35 crc kubenswrapper[4690]: I1203 13:36:35.583187 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 13:36:36 crc kubenswrapper[4690]: I1203 13:36:36.339725 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 13:36:36 crc kubenswrapper[4690]: I1203 13:36:36.935773 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 13:36:36 crc kubenswrapper[4690]: I1203 13:36:36.935821 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 13:36:37 crc kubenswrapper[4690]: I1203 13:36:37.951117 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3f0b2df1-b5ed-498c-b6dd-784a32f97b3f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.218:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:36:37 crc kubenswrapper[4690]: I1203 13:36:37.951155 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3f0b2df1-b5ed-498c-b6dd-784a32f97b3f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.218:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:36:39 crc kubenswrapper[4690]: I1203 13:36:39.232071 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 13:36:39 crc kubenswrapper[4690]: I1203 13:36:39.644846 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:36:39 crc kubenswrapper[4690]: I1203 13:36:39.645304 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:36:40 crc kubenswrapper[4690]: I1203 13:36:40.663108 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.219:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:36:40 crc kubenswrapper[4690]: I1203 13:36:40.663121 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.219:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:36:46 crc kubenswrapper[4690]: I1203 13:36:46.823775 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:36:46 crc kubenswrapper[4690]: I1203 13:36:46.824435 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:36:46 crc kubenswrapper[4690]: I1203 13:36:46.941543 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 13:36:46 crc kubenswrapper[4690]: I1203 13:36:46.943913 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 13:36:46 crc kubenswrapper[4690]: I1203 13:36:46.949393 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 13:36:47 crc kubenswrapper[4690]: I1203 13:36:47.442469 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 13:36:49 crc kubenswrapper[4690]: I1203 13:36:49.652929 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 13:36:49 crc kubenswrapper[4690]: I1203 13:36:49.653801 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 13:36:49 crc kubenswrapper[4690]: I1203 13:36:49.654028 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 13:36:49 crc kubenswrapper[4690]: I1203 13:36:49.663163 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 13:36:50 crc kubenswrapper[4690]: I1203 13:36:50.470635 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 13:36:50 crc kubenswrapper[4690]: I1203 13:36:50.478773 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 13:36:58 crc kubenswrapper[4690]: I1203 13:36:58.900066 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:37:00 crc kubenswrapper[4690]: I1203 13:37:00.032848 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:37:03 crc kubenswrapper[4690]: I1203 13:37:03.966088 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" containerName="rabbitmq" containerID="cri-o://e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1" gracePeriod=604795 Dec 03 13:37:04 crc kubenswrapper[4690]: I1203 13:37:04.899527 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerName="rabbitmq" containerID="cri-o://65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807" gracePeriod=604796 Dec 03 13:37:07 crc kubenswrapper[4690]: I1203 13:37:07.887260 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.113:5671: connect: connection refused" Dec 03 13:37:07 crc kubenswrapper[4690]: I1203 13:37:07.989905 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.114:5671: connect: connection refused" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.673703 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.716724 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-config-data\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.716953 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-server-conf\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.716984 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-erlang-cookie\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.717033 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.717073 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pz4h\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-kube-api-access-2pz4h\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.717139 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6173a433-90f9-4e48-8297-81767d5ce03e-erlang-cookie-secret\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.717230 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-tls\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.717315 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-confd\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.717424 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-plugins-conf\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.717486 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6173a433-90f9-4e48-8297-81767d5ce03e-pod-info\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.717543 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-plugins\") pod \"6173a433-90f9-4e48-8297-81767d5ce03e\" (UID: \"6173a433-90f9-4e48-8297-81767d5ce03e\") " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.718713 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.718781 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.719250 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.750238 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.750368 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6173a433-90f9-4e48-8297-81767d5ce03e-pod-info" (OuterVolumeSpecName: "pod-info") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.751895 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6173a433-90f9-4e48-8297-81767d5ce03e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.754394 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-kube-api-access-2pz4h" (OuterVolumeSpecName: "kube-api-access-2pz4h") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "kube-api-access-2pz4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.767255 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.774643 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-config-data" (OuterVolumeSpecName: "config-data") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827427 4690 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827503 4690 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6173a433-90f9-4e48-8297-81767d5ce03e-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827519 4690 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827528 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827537 4690 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827580 4690 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827590 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pz4h\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-kube-api-access-2pz4h\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827599 4690 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6173a433-90f9-4e48-8297-81767d5ce03e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.827608 4690 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.891139 4690 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.924289 4690 generic.go:334] "Generic (PLEG): container finished" podID="6173a433-90f9-4e48-8297-81767d5ce03e" containerID="e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1" exitCode=0 Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.924331 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6173a433-90f9-4e48-8297-81767d5ce03e","Type":"ContainerDied","Data":"e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1"} Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.924358 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6173a433-90f9-4e48-8297-81767d5ce03e","Type":"ContainerDied","Data":"cc86a2e8021c328d4a576e674888abb0efa60b2e6200e716205d5af605ecd167"} Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.924375 4690 scope.go:117] "RemoveContainer" containerID="e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.924541 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.938002 4690 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:10 crc kubenswrapper[4690]: I1203 13:37:10.956443 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-server-conf" (OuterVolumeSpecName: "server-conf") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.014225 4690 scope.go:117] "RemoveContainer" containerID="ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.039850 4690 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6173a433-90f9-4e48-8297-81767d5ce03e-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.070129 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6173a433-90f9-4e48-8297-81767d5ce03e" (UID: "6173a433-90f9-4e48-8297-81767d5ce03e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.093202 4690 scope.go:117] "RemoveContainer" containerID="e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1" Dec 03 13:37:11 crc kubenswrapper[4690]: E1203 13:37:11.101008 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1\": container with ID starting with e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1 not found: ID does not exist" containerID="e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.101078 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1"} err="failed to get container status \"e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1\": rpc error: code = NotFound desc = could not find container \"e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1\": container with ID starting with e7e2089ba2e2be0a36f11d7ec2d40ceca295ff6b6a62e98daf560347a50e3ff1 not found: ID does not exist" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.101117 4690 scope.go:117] "RemoveContainer" containerID="ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670" Dec 03 13:37:11 crc kubenswrapper[4690]: E1203 13:37:11.104368 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670\": container with ID starting with ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670 not found: ID does not exist" containerID="ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.104411 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670"} err="failed to get container status \"ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670\": rpc error: code = NotFound desc = could not find container \"ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670\": container with ID starting with ec9047233ddf35c41e1e47df400011407b5107a6c7f3711cadf70a0a97621670 not found: ID does not exist" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.142254 4690 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6173a433-90f9-4e48-8297-81767d5ce03e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.296957 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.319161 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.349540 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:37:11 crc kubenswrapper[4690]: E1203 13:37:11.350161 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" containerName="rabbitmq" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.350189 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" containerName="rabbitmq" Dec 03 13:37:11 crc kubenswrapper[4690]: E1203 13:37:11.350232 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" containerName="setup-container" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.350242 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" containerName="setup-container" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.350528 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" containerName="rabbitmq" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.352049 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.358332 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.358511 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-65kkw" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.358783 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.358977 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.359136 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.359313 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.359480 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.383094 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.550304 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-config-data\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.550750 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.550856 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg7zp\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-kube-api-access-pg7zp\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.551008 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.551184 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.551312 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.551386 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.551541 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.551674 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.551814 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.551894 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.653929 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654015 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654085 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-config-data\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654115 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654169 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg7zp\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-kube-api-access-pg7zp\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654199 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654251 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654297 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654323 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654375 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.654428 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.655473 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.655722 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-config-data\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.656187 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.656713 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.657020 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.657278 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.659880 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.660240 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.660403 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.660630 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.679788 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg7zp\" (UniqueName: \"kubernetes.io/projected/bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8-kube-api-access-pg7zp\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.701424 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8\") " pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.806160 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.886638 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-config-data\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.886713 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-plugins\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.886815 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.886860 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-erlang-cookie\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.886937 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-confd\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.886969 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5930069-12cd-4b0b-8aab-0f32ea9057a9-pod-info\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.886986 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-plugins-conf\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.887028 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-tls\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.887088 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btxts\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-kube-api-access-btxts\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.887116 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-server-conf\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.887174 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5930069-12cd-4b0b-8aab-0f32ea9057a9-erlang-cookie-secret\") pod \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\" (UID: \"d5930069-12cd-4b0b-8aab-0f32ea9057a9\") " Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.890037 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.890581 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.894739 4690 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.895129 4690 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.895545 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.914903 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.914964 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d5930069-12cd-4b0b-8aab-0f32ea9057a9-pod-info" (OuterVolumeSpecName: "pod-info") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.916434 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5930069-12cd-4b0b-8aab-0f32ea9057a9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.916637 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.943288 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-kube-api-access-btxts" (OuterVolumeSpecName: "kube-api-access-btxts") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "kube-api-access-btxts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.947350 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-config-data" (OuterVolumeSpecName: "config-data") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.974803 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.982221 4690 generic.go:334] "Generic (PLEG): container finished" podID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerID="65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807" exitCode=0 Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.982361 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5930069-12cd-4b0b-8aab-0f32ea9057a9","Type":"ContainerDied","Data":"65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807"} Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.982406 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.982427 4690 scope.go:117] "RemoveContainer" containerID="65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807" Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.982413 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5930069-12cd-4b0b-8aab-0f32ea9057a9","Type":"ContainerDied","Data":"8af360905b9e2c3a677ccd5b8c32087eb23ee9c42a552ff718e90680f92f716c"} Dec 03 13:37:11 crc kubenswrapper[4690]: I1203 13:37:11.986437 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-server-conf" (OuterVolumeSpecName: "server-conf") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.007923 4690 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.007978 4690 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5930069-12cd-4b0b-8aab-0f32ea9057a9-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.007993 4690 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.008006 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btxts\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-kube-api-access-btxts\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.008017 4690 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.008029 4690 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5930069-12cd-4b0b-8aab-0f32ea9057a9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.008040 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5930069-12cd-4b0b-8aab-0f32ea9057a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.008069 4690 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.048452 4690 scope.go:117] "RemoveContainer" containerID="4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.060426 4690 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.078411 4690 scope.go:117] "RemoveContainer" containerID="65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807" Dec 03 13:37:12 crc kubenswrapper[4690]: E1203 13:37:12.079769 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807\": container with ID starting with 65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807 not found: ID does not exist" containerID="65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.080047 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807"} err="failed to get container status \"65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807\": rpc error: code = NotFound desc = could not find container \"65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807\": container with ID starting with 65beec2982314ac0b132a3bfd151bd9b08dd07c824060cad105c5bf19c4a2807 not found: ID does not exist" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.080194 4690 scope.go:117] "RemoveContainer" containerID="4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928" Dec 03 13:37:12 crc kubenswrapper[4690]: E1203 13:37:12.081239 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928\": container with ID starting with 4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928 not found: ID does not exist" containerID="4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.081369 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928"} err="failed to get container status \"4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928\": rpc error: code = NotFound desc = could not find container \"4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928\": container with ID starting with 4e82ea0bd68bf084f0e1c266c3ea539aab4f4b79a7807c501e6d389a88183928 not found: ID does not exist" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.113483 4690 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.114115 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d5930069-12cd-4b0b-8aab-0f32ea9057a9" (UID: "d5930069-12cd-4b0b-8aab-0f32ea9057a9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.215335 4690 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5930069-12cd-4b0b-8aab-0f32ea9057a9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.342113 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6173a433-90f9-4e48-8297-81767d5ce03e" path="/var/lib/kubelet/pods/6173a433-90f9-4e48-8297-81767d5ce03e/volumes" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.359588 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.380435 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.399855 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:37:12 crc kubenswrapper[4690]: E1203 13:37:12.400573 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerName="setup-container" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.400624 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerName="setup-container" Dec 03 13:37:12 crc kubenswrapper[4690]: E1203 13:37:12.401620 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerName="rabbitmq" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.401639 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerName="rabbitmq" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.401936 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" containerName="rabbitmq" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.404074 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.410394 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.410715 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.410932 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.411114 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.411247 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-jptwb" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.411414 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.411430 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.411610 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.530164 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js5kl\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-kube-api-access-js5kl\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.530988 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.531019 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.531048 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.531267 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.531442 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.531539 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.531598 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.531852 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.531903 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.532052 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.585012 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.634595 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.635128 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.635152 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.635234 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.635979 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.636040 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js5kl\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-kube-api-access-js5kl\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.636195 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.636225 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.636265 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.636351 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.636436 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.636479 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.637393 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.637458 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.637780 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.637921 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.638218 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.641260 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.642325 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.643454 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.644177 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.656842 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js5kl\" (UniqueName: \"kubernetes.io/projected/8daef7dd-5b47-4f5c-8dc8-0da54d682e1d-kube-api-access-js5kl\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.688519 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:12 crc kubenswrapper[4690]: I1203 13:37:12.752732 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.020521 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8","Type":"ContainerStarted","Data":"5a8bb4036a73040ad240b4c5301dff2bdf9b2965c8bccc681bf161aac6aecd6d"} Dec 03 13:37:13 crc kubenswrapper[4690]: W1203 13:37:13.273654 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8daef7dd_5b47_4f5c_8dc8_0da54d682e1d.slice/crio-021d4b6fcb5aed83c3804cfe4ef86d36e0d8ba61e000f99d493160761e6d053a WatchSource:0}: Error finding container 021d4b6fcb5aed83c3804cfe4ef86d36e0d8ba61e000f99d493160761e6d053a: Status 404 returned error can't find the container with id 021d4b6fcb5aed83c3804cfe4ef86d36e0d8ba61e000f99d493160761e6d053a Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.276116 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.410169 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-zz7jf"] Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.412445 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.415010 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.446915 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-zz7jf"] Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.456440 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.456528 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.456593 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.456656 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhnfx\" (UniqueName: \"kubernetes.io/projected/c79ffe20-44af-4607-b728-ada656f3c9f6-kube-api-access-qhnfx\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.456762 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.456803 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.456904 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-config\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.559040 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.559492 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhnfx\" (UniqueName: \"kubernetes.io/projected/c79ffe20-44af-4607-b728-ada656f3c9f6-kube-api-access-qhnfx\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.559603 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.559645 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.559724 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-config\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.559770 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.559815 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.560296 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.560593 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.561016 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.561247 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-config\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.561810 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.562124 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.585976 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhnfx\" (UniqueName: \"kubernetes.io/projected/c79ffe20-44af-4607-b728-ada656f3c9f6-kube-api-access-qhnfx\") pod \"dnsmasq-dns-79bd4cc8c9-zz7jf\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:13 crc kubenswrapper[4690]: I1203 13:37:13.801842 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:14 crc kubenswrapper[4690]: I1203 13:37:14.082601 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d","Type":"ContainerStarted","Data":"021d4b6fcb5aed83c3804cfe4ef86d36e0d8ba61e000f99d493160761e6d053a"} Dec 03 13:37:14 crc kubenswrapper[4690]: I1203 13:37:14.348973 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5930069-12cd-4b0b-8aab-0f32ea9057a9" path="/var/lib/kubelet/pods/d5930069-12cd-4b0b-8aab-0f32ea9057a9/volumes" Dec 03 13:37:14 crc kubenswrapper[4690]: I1203 13:37:14.436021 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-zz7jf"] Dec 03 13:37:14 crc kubenswrapper[4690]: W1203 13:37:14.561741 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc79ffe20_44af_4607_b728_ada656f3c9f6.slice/crio-5f3cb812459cb91a883a7ed9a5100288006048f19283b19678faf2c4c3538cdb WatchSource:0}: Error finding container 5f3cb812459cb91a883a7ed9a5100288006048f19283b19678faf2c4c3538cdb: Status 404 returned error can't find the container with id 5f3cb812459cb91a883a7ed9a5100288006048f19283b19678faf2c4c3538cdb Dec 03 13:37:15 crc kubenswrapper[4690]: I1203 13:37:15.093982 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8","Type":"ContainerStarted","Data":"1b10f46810452360b17bdfe31203c1d429c3c85c8b3e0964d4e4e62a4f1b01e6"} Dec 03 13:37:15 crc kubenswrapper[4690]: I1203 13:37:15.095885 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" event={"ID":"c79ffe20-44af-4607-b728-ada656f3c9f6","Type":"ContainerStarted","Data":"5f3cb812459cb91a883a7ed9a5100288006048f19283b19678faf2c4c3538cdb"} Dec 03 13:37:16 crc kubenswrapper[4690]: I1203 13:37:16.111327 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d","Type":"ContainerStarted","Data":"c5fe2a0ec3e9b8c12efc69ad20e88f47faa2f774d88d8456a1d2221624fd99d6"} Dec 03 13:37:16 crc kubenswrapper[4690]: I1203 13:37:16.114745 4690 generic.go:334] "Generic (PLEG): container finished" podID="c79ffe20-44af-4607-b728-ada656f3c9f6" containerID="19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8" exitCode=0 Dec 03 13:37:16 crc kubenswrapper[4690]: I1203 13:37:16.114830 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" event={"ID":"c79ffe20-44af-4607-b728-ada656f3c9f6","Type":"ContainerDied","Data":"19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8"} Dec 03 13:37:16 crc kubenswrapper[4690]: I1203 13:37:16.823591 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:37:16 crc kubenswrapper[4690]: I1203 13:37:16.823946 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:37:16 crc kubenswrapper[4690]: I1203 13:37:16.824016 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:37:16 crc kubenswrapper[4690]: I1203 13:37:16.824947 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:37:16 crc kubenswrapper[4690]: I1203 13:37:16.825018 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" gracePeriod=600 Dec 03 13:37:16 crc kubenswrapper[4690]: E1203 13:37:16.963947 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:37:17 crc kubenswrapper[4690]: I1203 13:37:17.128556 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" event={"ID":"c79ffe20-44af-4607-b728-ada656f3c9f6","Type":"ContainerStarted","Data":"9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c"} Dec 03 13:37:17 crc kubenswrapper[4690]: I1203 13:37:17.130443 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:17 crc kubenswrapper[4690]: I1203 13:37:17.133475 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" exitCode=0 Dec 03 13:37:17 crc kubenswrapper[4690]: I1203 13:37:17.134146 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2"} Dec 03 13:37:17 crc kubenswrapper[4690]: I1203 13:37:17.134186 4690 scope.go:117] "RemoveContainer" containerID="d13f0eb20e5e81edaffaa899e1251f34d5ba5b9191f18b642a8f24f971ecc07d" Dec 03 13:37:17 crc kubenswrapper[4690]: I1203 13:37:17.134546 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:37:17 crc kubenswrapper[4690]: E1203 13:37:17.134788 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:37:17 crc kubenswrapper[4690]: I1203 13:37:17.165743 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" podStartSLOduration=4.165713086 podStartE2EDuration="4.165713086s" podCreationTimestamp="2025-12-03 13:37:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:37:17.160187502 +0000 UTC m=+1683.141107925" watchObservedRunningTime="2025-12-03 13:37:17.165713086 +0000 UTC m=+1683.146633519" Dec 03 13:37:23 crc kubenswrapper[4690]: I1203 13:37:23.804309 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:23 crc kubenswrapper[4690]: I1203 13:37:23.903667 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dphf7"] Dec 03 13:37:23 crc kubenswrapper[4690]: I1203 13:37:23.903964 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" podUID="0303b53b-3e79-40cf-8769-65addeca1be9" containerName="dnsmasq-dns" containerID="cri-o://7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd" gracePeriod=10 Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.058951 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-wn8bx"] Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.062044 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.075141 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-wn8bx"] Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.127102 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf78x\" (UniqueName: \"kubernetes.io/projected/fa48f557-98dd-49ca-a4a6-89b196b77b15-kube-api-access-qf78x\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.127193 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-dns-svc\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.127262 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.127287 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.127572 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.127794 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.127944 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-config\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.229686 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.229773 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.229819 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-config\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.229887 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf78x\" (UniqueName: \"kubernetes.io/projected/fa48f557-98dd-49ca-a4a6-89b196b77b15-kube-api-access-qf78x\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.229921 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-dns-svc\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.229947 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.229967 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.230815 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.231483 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.231638 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-config\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.232151 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.232205 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-dns-svc\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.232416 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa48f557-98dd-49ca-a4a6-89b196b77b15-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.254236 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf78x\" (UniqueName: \"kubernetes.io/projected/fa48f557-98dd-49ca-a4a6-89b196b77b15-kube-api-access-qf78x\") pod \"dnsmasq-dns-55478c4467-wn8bx\" (UID: \"fa48f557-98dd-49ca-a4a6-89b196b77b15\") " pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.255850 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" podUID="0303b53b-3e79-40cf-8769-65addeca1be9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.213:5353: connect: connection refused" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.406751 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:24 crc kubenswrapper[4690]: I1203 13:37:24.918965 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-wn8bx"] Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.214716 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.227118 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-wn8bx" event={"ID":"fa48f557-98dd-49ca-a4a6-89b196b77b15","Type":"ContainerStarted","Data":"f3fb17bd1ede2a5f35b0cd2ccc147c21e9c4e3e61898e68d20b97e1d09d3f49a"} Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.229786 4690 generic.go:334] "Generic (PLEG): container finished" podID="0303b53b-3e79-40cf-8769-65addeca1be9" containerID="7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd" exitCode=0 Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.229855 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.229858 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" event={"ID":"0303b53b-3e79-40cf-8769-65addeca1be9","Type":"ContainerDied","Data":"7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd"} Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.230151 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-dphf7" event={"ID":"0303b53b-3e79-40cf-8769-65addeca1be9","Type":"ContainerDied","Data":"7f9c48faff043602d087a1d223c2d5f0f718d37c75629ada4ed6bfeccea1a9e6"} Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.230181 4690 scope.go:117] "RemoveContainer" containerID="7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.261944 4690 scope.go:117] "RemoveContainer" containerID="8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.262774 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h44kk\" (UniqueName: \"kubernetes.io/projected/0303b53b-3e79-40cf-8769-65addeca1be9-kube-api-access-h44kk\") pod \"0303b53b-3e79-40cf-8769-65addeca1be9\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.262822 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-sb\") pod \"0303b53b-3e79-40cf-8769-65addeca1be9\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.262982 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-svc\") pod \"0303b53b-3e79-40cf-8769-65addeca1be9\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.263064 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-swift-storage-0\") pod \"0303b53b-3e79-40cf-8769-65addeca1be9\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.263127 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-nb\") pod \"0303b53b-3e79-40cf-8769-65addeca1be9\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.263187 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-config\") pod \"0303b53b-3e79-40cf-8769-65addeca1be9\" (UID: \"0303b53b-3e79-40cf-8769-65addeca1be9\") " Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.269100 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0303b53b-3e79-40cf-8769-65addeca1be9-kube-api-access-h44kk" (OuterVolumeSpecName: "kube-api-access-h44kk") pod "0303b53b-3e79-40cf-8769-65addeca1be9" (UID: "0303b53b-3e79-40cf-8769-65addeca1be9"). InnerVolumeSpecName "kube-api-access-h44kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.370422 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h44kk\" (UniqueName: \"kubernetes.io/projected/0303b53b-3e79-40cf-8769-65addeca1be9-kube-api-access-h44kk\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.377915 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0303b53b-3e79-40cf-8769-65addeca1be9" (UID: "0303b53b-3e79-40cf-8769-65addeca1be9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.381457 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-config" (OuterVolumeSpecName: "config") pod "0303b53b-3e79-40cf-8769-65addeca1be9" (UID: "0303b53b-3e79-40cf-8769-65addeca1be9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.392404 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0303b53b-3e79-40cf-8769-65addeca1be9" (UID: "0303b53b-3e79-40cf-8769-65addeca1be9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.395352 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0303b53b-3e79-40cf-8769-65addeca1be9" (UID: "0303b53b-3e79-40cf-8769-65addeca1be9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.400013 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0303b53b-3e79-40cf-8769-65addeca1be9" (UID: "0303b53b-3e79-40cf-8769-65addeca1be9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.438578 4690 scope.go:117] "RemoveContainer" containerID="7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd" Dec 03 13:37:25 crc kubenswrapper[4690]: E1203 13:37:25.439391 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd\": container with ID starting with 7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd not found: ID does not exist" containerID="7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.439447 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd"} err="failed to get container status \"7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd\": rpc error: code = NotFound desc = could not find container \"7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd\": container with ID starting with 7c04090eae2e71692fa949c799f361d5edde497a0304e189852669ce8eff2ccd not found: ID does not exist" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.439484 4690 scope.go:117] "RemoveContainer" containerID="8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854" Dec 03 13:37:25 crc kubenswrapper[4690]: E1203 13:37:25.439978 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854\": container with ID starting with 8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854 not found: ID does not exist" containerID="8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.440006 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854"} err="failed to get container status \"8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854\": rpc error: code = NotFound desc = could not find container \"8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854\": container with ID starting with 8e94b36077a945917cb56c250807df2eab0612f904d0b4c8a8937a1d0dffd854 not found: ID does not exist" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.475162 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.475201 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.475213 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.475224 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.475235 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0303b53b-3e79-40cf-8769-65addeca1be9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.570458 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dphf7"] Dec 03 13:37:25 crc kubenswrapper[4690]: I1203 13:37:25.582282 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-dphf7"] Dec 03 13:37:26 crc kubenswrapper[4690]: I1203 13:37:26.240157 4690 generic.go:334] "Generic (PLEG): container finished" podID="fa48f557-98dd-49ca-a4a6-89b196b77b15" containerID="93e7f7b75fc6554178e0b291997dbfa9197f4b0338e9cc2d7aaf81f43c9d32a9" exitCode=0 Dec 03 13:37:26 crc kubenswrapper[4690]: I1203 13:37:26.240227 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-wn8bx" event={"ID":"fa48f557-98dd-49ca-a4a6-89b196b77b15","Type":"ContainerDied","Data":"93e7f7b75fc6554178e0b291997dbfa9197f4b0338e9cc2d7aaf81f43c9d32a9"} Dec 03 13:37:26 crc kubenswrapper[4690]: I1203 13:37:26.328191 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0303b53b-3e79-40cf-8769-65addeca1be9" path="/var/lib/kubelet/pods/0303b53b-3e79-40cf-8769-65addeca1be9/volumes" Dec 03 13:37:27 crc kubenswrapper[4690]: I1203 13:37:27.257734 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-wn8bx" event={"ID":"fa48f557-98dd-49ca-a4a6-89b196b77b15","Type":"ContainerStarted","Data":"27720dead7221a5cc48991a7eccfcc69d6516b106d70d12fc246e644f6de88e6"} Dec 03 13:37:27 crc kubenswrapper[4690]: I1203 13:37:27.258100 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:27 crc kubenswrapper[4690]: I1203 13:37:27.295120 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-wn8bx" podStartSLOduration=3.295098507 podStartE2EDuration="3.295098507s" podCreationTimestamp="2025-12-03 13:37:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:37:27.277249502 +0000 UTC m=+1693.258169945" watchObservedRunningTime="2025-12-03 13:37:27.295098507 +0000 UTC m=+1693.276018940" Dec 03 13:37:28 crc kubenswrapper[4690]: I1203 13:37:28.314945 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:37:28 crc kubenswrapper[4690]: E1203 13:37:28.315649 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:37:34 crc kubenswrapper[4690]: I1203 13:37:34.408419 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-wn8bx" Dec 03 13:37:34 crc kubenswrapper[4690]: I1203 13:37:34.488667 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-zz7jf"] Dec 03 13:37:34 crc kubenswrapper[4690]: I1203 13:37:34.489064 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" podUID="c79ffe20-44af-4607-b728-ada656f3c9f6" containerName="dnsmasq-dns" containerID="cri-o://9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c" gracePeriod=10 Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.288422 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.394100 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhnfx\" (UniqueName: \"kubernetes.io/projected/c79ffe20-44af-4607-b728-ada656f3c9f6-kube-api-access-qhnfx\") pod \"c79ffe20-44af-4607-b728-ada656f3c9f6\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.394160 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-sb\") pod \"c79ffe20-44af-4607-b728-ada656f3c9f6\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.394206 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-config\") pod \"c79ffe20-44af-4607-b728-ada656f3c9f6\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.394230 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-openstack-edpm-ipam\") pod \"c79ffe20-44af-4607-b728-ada656f3c9f6\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.394255 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-swift-storage-0\") pod \"c79ffe20-44af-4607-b728-ada656f3c9f6\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.394362 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-nb\") pod \"c79ffe20-44af-4607-b728-ada656f3c9f6\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.394450 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-svc\") pod \"c79ffe20-44af-4607-b728-ada656f3c9f6\" (UID: \"c79ffe20-44af-4607-b728-ada656f3c9f6\") " Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.403984 4690 generic.go:334] "Generic (PLEG): container finished" podID="c79ffe20-44af-4607-b728-ada656f3c9f6" containerID="9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c" exitCode=0 Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.404067 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" event={"ID":"c79ffe20-44af-4607-b728-ada656f3c9f6","Type":"ContainerDied","Data":"9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c"} Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.404181 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" event={"ID":"c79ffe20-44af-4607-b728-ada656f3c9f6","Type":"ContainerDied","Data":"5f3cb812459cb91a883a7ed9a5100288006048f19283b19678faf2c4c3538cdb"} Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.404206 4690 scope.go:117] "RemoveContainer" containerID="9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.404463 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-zz7jf" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.416166 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c79ffe20-44af-4607-b728-ada656f3c9f6-kube-api-access-qhnfx" (OuterVolumeSpecName: "kube-api-access-qhnfx") pod "c79ffe20-44af-4607-b728-ada656f3c9f6" (UID: "c79ffe20-44af-4607-b728-ada656f3c9f6"). InnerVolumeSpecName "kube-api-access-qhnfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.471546 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c79ffe20-44af-4607-b728-ada656f3c9f6" (UID: "c79ffe20-44af-4607-b728-ada656f3c9f6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.476369 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "c79ffe20-44af-4607-b728-ada656f3c9f6" (UID: "c79ffe20-44af-4607-b728-ada656f3c9f6"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.479736 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c79ffe20-44af-4607-b728-ada656f3c9f6" (UID: "c79ffe20-44af-4607-b728-ada656f3c9f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.482171 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c79ffe20-44af-4607-b728-ada656f3c9f6" (UID: "c79ffe20-44af-4607-b728-ada656f3c9f6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.483587 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c79ffe20-44af-4607-b728-ada656f3c9f6" (UID: "c79ffe20-44af-4607-b728-ada656f3c9f6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.492216 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-config" (OuterVolumeSpecName: "config") pod "c79ffe20-44af-4607-b728-ada656f3c9f6" (UID: "c79ffe20-44af-4607-b728-ada656f3c9f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.496976 4690 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.497022 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.497042 4690 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.497058 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.497072 4690 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.497083 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhnfx\" (UniqueName: \"kubernetes.io/projected/c79ffe20-44af-4607-b728-ada656f3c9f6-kube-api-access-qhnfx\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.497093 4690 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79ffe20-44af-4607-b728-ada656f3c9f6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.544413 4690 scope.go:117] "RemoveContainer" containerID="19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.573113 4690 scope.go:117] "RemoveContainer" containerID="9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c" Dec 03 13:37:35 crc kubenswrapper[4690]: E1203 13:37:35.573611 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c\": container with ID starting with 9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c not found: ID does not exist" containerID="9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.573660 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c"} err="failed to get container status \"9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c\": rpc error: code = NotFound desc = could not find container \"9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c\": container with ID starting with 9f8286d53009fc2e2a0dd7090658bdf316ff1b5e5fdf333073424b72e0582a1c not found: ID does not exist" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.573692 4690 scope.go:117] "RemoveContainer" containerID="19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8" Dec 03 13:37:35 crc kubenswrapper[4690]: E1203 13:37:35.574089 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8\": container with ID starting with 19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8 not found: ID does not exist" containerID="19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.574131 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8"} err="failed to get container status \"19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8\": rpc error: code = NotFound desc = could not find container \"19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8\": container with ID starting with 19289d1022e5db79df13a0941c21157b1e8e7cfc98bc1046f97194713031fba8 not found: ID does not exist" Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.738645 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-zz7jf"] Dec 03 13:37:35 crc kubenswrapper[4690]: I1203 13:37:35.749002 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-zz7jf"] Dec 03 13:37:36 crc kubenswrapper[4690]: I1203 13:37:36.328192 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c79ffe20-44af-4607-b728-ada656f3c9f6" path="/var/lib/kubelet/pods/c79ffe20-44af-4607-b728-ada656f3c9f6/volumes" Dec 03 13:37:39 crc kubenswrapper[4690]: I1203 13:37:39.588799 4690 scope.go:117] "RemoveContainer" containerID="6aa3703018330b11655b823fe6b157d787f28a70f379fdebf89ef23a9401accf" Dec 03 13:37:42 crc kubenswrapper[4690]: I1203 13:37:42.315156 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:37:42 crc kubenswrapper[4690]: E1203 13:37:42.316009 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.427107 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm"] Dec 03 13:37:47 crc kubenswrapper[4690]: E1203 13:37:47.429592 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79ffe20-44af-4607-b728-ada656f3c9f6" containerName="init" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.429855 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79ffe20-44af-4607-b728-ada656f3c9f6" containerName="init" Dec 03 13:37:47 crc kubenswrapper[4690]: E1203 13:37:47.430116 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0303b53b-3e79-40cf-8769-65addeca1be9" containerName="dnsmasq-dns" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.430194 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="0303b53b-3e79-40cf-8769-65addeca1be9" containerName="dnsmasq-dns" Dec 03 13:37:47 crc kubenswrapper[4690]: E1203 13:37:47.430271 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79ffe20-44af-4607-b728-ada656f3c9f6" containerName="dnsmasq-dns" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.430332 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79ffe20-44af-4607-b728-ada656f3c9f6" containerName="dnsmasq-dns" Dec 03 13:37:47 crc kubenswrapper[4690]: E1203 13:37:47.430403 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0303b53b-3e79-40cf-8769-65addeca1be9" containerName="init" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.430465 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="0303b53b-3e79-40cf-8769-65addeca1be9" containerName="init" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.430767 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="0303b53b-3e79-40cf-8769-65addeca1be9" containerName="dnsmasq-dns" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.430847 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79ffe20-44af-4607-b728-ada656f3c9f6" containerName="dnsmasq-dns" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.431791 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.433794 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.434270 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.434574 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.436419 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.440411 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm"] Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.481601 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.481694 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.482028 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.482157 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dthrb\" (UniqueName: \"kubernetes.io/projected/154bbd71-1585-49b0-84ab-3d11d49c99e6-kube-api-access-dthrb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.548523 4690 generic.go:334] "Generic (PLEG): container finished" podID="bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8" containerID="1b10f46810452360b17bdfe31203c1d429c3c85c8b3e0964d4e4e62a4f1b01e6" exitCode=0 Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.548572 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8","Type":"ContainerDied","Data":"1b10f46810452360b17bdfe31203c1d429c3c85c8b3e0964d4e4e62a4f1b01e6"} Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.587519 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dthrb\" (UniqueName: \"kubernetes.io/projected/154bbd71-1585-49b0-84ab-3d11d49c99e6-kube-api-access-dthrb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.587597 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.587670 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.587793 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.592915 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.593183 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.593943 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.607847 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dthrb\" (UniqueName: \"kubernetes.io/projected/154bbd71-1585-49b0-84ab-3d11d49c99e6-kube-api-access-dthrb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:47 crc kubenswrapper[4690]: I1203 13:37:47.759539 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:37:48 crc kubenswrapper[4690]: I1203 13:37:48.302539 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm"] Dec 03 13:37:48 crc kubenswrapper[4690]: W1203 13:37:48.302791 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod154bbd71_1585_49b0_84ab_3d11d49c99e6.slice/crio-38b7c9da9bdc32c214a7dbf8b7e04767727f756e034df4bcbe1f89d5632ddcef WatchSource:0}: Error finding container 38b7c9da9bdc32c214a7dbf8b7e04767727f756e034df4bcbe1f89d5632ddcef: Status 404 returned error can't find the container with id 38b7c9da9bdc32c214a7dbf8b7e04767727f756e034df4bcbe1f89d5632ddcef Dec 03 13:37:48 crc kubenswrapper[4690]: I1203 13:37:48.305709 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:37:48 crc kubenswrapper[4690]: I1203 13:37:48.561606 4690 generic.go:334] "Generic (PLEG): container finished" podID="8daef7dd-5b47-4f5c-8dc8-0da54d682e1d" containerID="c5fe2a0ec3e9b8c12efc69ad20e88f47faa2f774d88d8456a1d2221624fd99d6" exitCode=0 Dec 03 13:37:48 crc kubenswrapper[4690]: I1203 13:37:48.561702 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d","Type":"ContainerDied","Data":"c5fe2a0ec3e9b8c12efc69ad20e88f47faa2f774d88d8456a1d2221624fd99d6"} Dec 03 13:37:48 crc kubenswrapper[4690]: I1203 13:37:48.564149 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" event={"ID":"154bbd71-1585-49b0-84ab-3d11d49c99e6","Type":"ContainerStarted","Data":"38b7c9da9bdc32c214a7dbf8b7e04767727f756e034df4bcbe1f89d5632ddcef"} Dec 03 13:37:49 crc kubenswrapper[4690]: I1203 13:37:49.575806 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8","Type":"ContainerStarted","Data":"5e283c2c300ce12f44915d228de26d1e8eb35c3aeaa3c503ba3263b111567733"} Dec 03 13:37:49 crc kubenswrapper[4690]: I1203 13:37:49.578551 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8daef7dd-5b47-4f5c-8dc8-0da54d682e1d","Type":"ContainerStarted","Data":"d9cb8be728f5165bea6b72d0781345b98b2b17cfd7b912da345f665870add08f"} Dec 03 13:37:51 crc kubenswrapper[4690]: I1203 13:37:51.600386 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 13:37:51 crc kubenswrapper[4690]: I1203 13:37:51.635983 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.635961466 podStartE2EDuration="40.635961466s" podCreationTimestamp="2025-12-03 13:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:37:51.623527832 +0000 UTC m=+1717.604448285" watchObservedRunningTime="2025-12-03 13:37:51.635961466 +0000 UTC m=+1717.616881899" Dec 03 13:37:51 crc kubenswrapper[4690]: I1203 13:37:51.664011 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.663984897 podStartE2EDuration="39.663984897s" podCreationTimestamp="2025-12-03 13:37:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:37:51.650970098 +0000 UTC m=+1717.631890541" watchObservedRunningTime="2025-12-03 13:37:51.663984897 +0000 UTC m=+1717.644905330" Dec 03 13:37:52 crc kubenswrapper[4690]: I1203 13:37:52.753696 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:37:53 crc kubenswrapper[4690]: I1203 13:37:53.315303 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:37:53 crc kubenswrapper[4690]: E1203 13:37:53.315656 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:38:01 crc kubenswrapper[4690]: I1203 13:38:01.979216 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.220:5671: connect: connection refused" Dec 03 13:38:02 crc kubenswrapper[4690]: I1203 13:38:02.755679 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="8daef7dd-5b47-4f5c-8dc8-0da54d682e1d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.221:5671: connect: connection refused" Dec 03 13:38:04 crc kubenswrapper[4690]: I1203 13:38:04.321785 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:38:04 crc kubenswrapper[4690]: E1203 13:38:04.322642 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:38:11 crc kubenswrapper[4690]: I1203 13:38:11.975916 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.220:5671: connect: connection refused" Dec 03 13:38:12 crc kubenswrapper[4690]: I1203 13:38:12.753917 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="8daef7dd-5b47-4f5c-8dc8-0da54d682e1d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.221:5671: connect: connection refused" Dec 03 13:38:17 crc kubenswrapper[4690]: I1203 13:38:17.316595 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:38:17 crc kubenswrapper[4690]: E1203 13:38:17.317663 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:38:21 crc kubenswrapper[4690]: I1203 13:38:21.976237 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.220:5671: connect: connection refused" Dec 03 13:38:22 crc kubenswrapper[4690]: I1203 13:38:22.753992 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="8daef7dd-5b47-4f5c-8dc8-0da54d682e1d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.221:5671: connect: connection refused" Dec 03 13:38:24 crc kubenswrapper[4690]: E1203 13:38:24.628257 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Dec 03 13:38:24 crc kubenswrapper[4690]: E1203 13:38:24.628933 4690 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 03 13:38:24 crc kubenswrapper[4690]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Dec 03 13:38:24 crc kubenswrapper[4690]: - hosts: all Dec 03 13:38:24 crc kubenswrapper[4690]: strategy: linear Dec 03 13:38:24 crc kubenswrapper[4690]: tasks: Dec 03 13:38:24 crc kubenswrapper[4690]: - name: Enable podified-repos Dec 03 13:38:24 crc kubenswrapper[4690]: become: true Dec 03 13:38:24 crc kubenswrapper[4690]: ansible.builtin.shell: | Dec 03 13:38:24 crc kubenswrapper[4690]: set -euxo pipefail Dec 03 13:38:24 crc kubenswrapper[4690]: pushd /var/tmp Dec 03 13:38:24 crc kubenswrapper[4690]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Dec 03 13:38:24 crc kubenswrapper[4690]: pushd repo-setup-main Dec 03 13:38:24 crc kubenswrapper[4690]: python3 -m venv ./venv Dec 03 13:38:24 crc kubenswrapper[4690]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Dec 03 13:38:24 crc kubenswrapper[4690]: ./venv/bin/repo-setup current-podified -b antelope Dec 03 13:38:24 crc kubenswrapper[4690]: popd Dec 03 13:38:24 crc kubenswrapper[4690]: rm -rf repo-setup-main Dec 03 13:38:24 crc kubenswrapper[4690]: Dec 03 13:38:24 crc kubenswrapper[4690]: Dec 03 13:38:24 crc kubenswrapper[4690]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Dec 03 13:38:24 crc kubenswrapper[4690]: edpm_override_hosts: openstack-edpm-ipam Dec 03 13:38:24 crc kubenswrapper[4690]: edpm_service_type: repo-setup Dec 03 13:38:24 crc kubenswrapper[4690]: Dec 03 13:38:24 crc kubenswrapper[4690]: Dec 03 13:38:24 crc kubenswrapper[4690]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dthrb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm_openstack(154bbd71-1585-49b0-84ab-3d11d49c99e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Dec 03 13:38:24 crc kubenswrapper[4690]: > logger="UnhandledError" Dec 03 13:38:24 crc kubenswrapper[4690]: E1203 13:38:24.630688 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" podUID="154bbd71-1585-49b0-84ab-3d11d49c99e6" Dec 03 13:38:24 crc kubenswrapper[4690]: E1203 13:38:24.985684 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" podUID="154bbd71-1585-49b0-84ab-3d11d49c99e6" Dec 03 13:38:30 crc kubenswrapper[4690]: I1203 13:38:30.315305 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:38:30 crc kubenswrapper[4690]: E1203 13:38:30.318179 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:38:31 crc kubenswrapper[4690]: I1203 13:38:31.976385 4690 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.220:5671: connect: connection refused" Dec 03 13:38:32 crc kubenswrapper[4690]: I1203 13:38:32.754927 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:38:37 crc kubenswrapper[4690]: I1203 13:38:37.419561 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:38:38 crc kubenswrapper[4690]: I1203 13:38:38.161333 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" event={"ID":"154bbd71-1585-49b0-84ab-3d11d49c99e6","Type":"ContainerStarted","Data":"a1fb0de7073dba4783ad9a06843bd141a8a0086f7d1b64f1aaf7458f0eea3e8b"} Dec 03 13:38:38 crc kubenswrapper[4690]: I1203 13:38:38.184937 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" podStartSLOduration=2.07458611 podStartE2EDuration="51.184908422s" podCreationTimestamp="2025-12-03 13:37:47 +0000 UTC" firstStartedPulling="2025-12-03 13:37:48.305481201 +0000 UTC m=+1714.286401634" lastFinishedPulling="2025-12-03 13:38:37.415803523 +0000 UTC m=+1763.396723946" observedRunningTime="2025-12-03 13:38:38.1755208 +0000 UTC m=+1764.156441233" watchObservedRunningTime="2025-12-03 13:38:38.184908422 +0000 UTC m=+1764.165828885" Dec 03 13:38:39 crc kubenswrapper[4690]: I1203 13:38:39.744467 4690 scope.go:117] "RemoveContainer" containerID="72e74369053341505d34978936a6f7ddf7df8ec8c983484863ea6f3ec0a362ea" Dec 03 13:38:41 crc kubenswrapper[4690]: I1203 13:38:41.977089 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 13:38:42 crc kubenswrapper[4690]: I1203 13:38:42.314585 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:38:42 crc kubenswrapper[4690]: E1203 13:38:42.315716 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:38:53 crc kubenswrapper[4690]: I1203 13:38:53.315095 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:38:53 crc kubenswrapper[4690]: E1203 13:38:53.316150 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:39:04 crc kubenswrapper[4690]: I1203 13:39:04.324309 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:39:04 crc kubenswrapper[4690]: E1203 13:39:04.325206 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:39:19 crc kubenswrapper[4690]: I1203 13:39:19.315155 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:39:19 crc kubenswrapper[4690]: E1203 13:39:19.316023 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:39:31 crc kubenswrapper[4690]: I1203 13:39:31.315535 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:39:31 crc kubenswrapper[4690]: E1203 13:39:31.316536 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:39:38 crc kubenswrapper[4690]: I1203 13:39:38.913960 4690 generic.go:334] "Generic (PLEG): container finished" podID="154bbd71-1585-49b0-84ab-3d11d49c99e6" containerID="a1fb0de7073dba4783ad9a06843bd141a8a0086f7d1b64f1aaf7458f0eea3e8b" exitCode=0 Dec 03 13:39:38 crc kubenswrapper[4690]: I1203 13:39:38.914041 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" event={"ID":"154bbd71-1585-49b0-84ab-3d11d49c99e6","Type":"ContainerDied","Data":"a1fb0de7073dba4783ad9a06843bd141a8a0086f7d1b64f1aaf7458f0eea3e8b"} Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.409472 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.446259 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-repo-setup-combined-ca-bundle\") pod \"154bbd71-1585-49b0-84ab-3d11d49c99e6\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.446382 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dthrb\" (UniqueName: \"kubernetes.io/projected/154bbd71-1585-49b0-84ab-3d11d49c99e6-kube-api-access-dthrb\") pod \"154bbd71-1585-49b0-84ab-3d11d49c99e6\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.446545 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-inventory\") pod \"154bbd71-1585-49b0-84ab-3d11d49c99e6\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.446888 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-ssh-key\") pod \"154bbd71-1585-49b0-84ab-3d11d49c99e6\" (UID: \"154bbd71-1585-49b0-84ab-3d11d49c99e6\") " Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.454521 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "154bbd71-1585-49b0-84ab-3d11d49c99e6" (UID: "154bbd71-1585-49b0-84ab-3d11d49c99e6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.455201 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/154bbd71-1585-49b0-84ab-3d11d49c99e6-kube-api-access-dthrb" (OuterVolumeSpecName: "kube-api-access-dthrb") pod "154bbd71-1585-49b0-84ab-3d11d49c99e6" (UID: "154bbd71-1585-49b0-84ab-3d11d49c99e6"). InnerVolumeSpecName "kube-api-access-dthrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.480763 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-inventory" (OuterVolumeSpecName: "inventory") pod "154bbd71-1585-49b0-84ab-3d11d49c99e6" (UID: "154bbd71-1585-49b0-84ab-3d11d49c99e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.491017 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "154bbd71-1585-49b0-84ab-3d11d49c99e6" (UID: "154bbd71-1585-49b0-84ab-3d11d49c99e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.550073 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.550117 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.550130 4690 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154bbd71-1585-49b0-84ab-3d11d49c99e6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.550142 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dthrb\" (UniqueName: \"kubernetes.io/projected/154bbd71-1585-49b0-84ab-3d11d49c99e6-kube-api-access-dthrb\") on node \"crc\" DevicePath \"\"" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.935786 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" event={"ID":"154bbd71-1585-49b0-84ab-3d11d49c99e6","Type":"ContainerDied","Data":"38b7c9da9bdc32c214a7dbf8b7e04767727f756e034df4bcbe1f89d5632ddcef"} Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.935835 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38b7c9da9bdc32c214a7dbf8b7e04767727f756e034df4bcbe1f89d5632ddcef" Dec 03 13:39:40 crc kubenswrapper[4690]: I1203 13:39:40.935856 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.039079 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x"] Dec 03 13:39:41 crc kubenswrapper[4690]: E1203 13:39:41.039699 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154bbd71-1585-49b0-84ab-3d11d49c99e6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.039726 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="154bbd71-1585-49b0-84ab-3d11d49c99e6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.039974 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="154bbd71-1585-49b0-84ab-3d11d49c99e6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.040992 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.044006 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.044227 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.045886 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.046264 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.053396 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x"] Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.060172 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c92p\" (UniqueName: \"kubernetes.io/projected/d4625caf-d841-4682-ae69-e7d2d90ef220-kube-api-access-4c92p\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.060273 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.060474 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.162412 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c92p\" (UniqueName: \"kubernetes.io/projected/d4625caf-d841-4682-ae69-e7d2d90ef220-kube-api-access-4c92p\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.162541 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.162753 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.166766 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.167295 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.186793 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c92p\" (UniqueName: \"kubernetes.io/projected/d4625caf-d841-4682-ae69-e7d2d90ef220-kube-api-access-4c92p\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f924x\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.363038 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.927482 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x"] Dec 03 13:39:41 crc kubenswrapper[4690]: I1203 13:39:41.962146 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" event={"ID":"d4625caf-d841-4682-ae69-e7d2d90ef220","Type":"ContainerStarted","Data":"8318867fd1e4d6b5fa5469224284d9c4a5d377a35529d8efa614f08720eaacd1"} Dec 03 13:39:42 crc kubenswrapper[4690]: I1203 13:39:42.314927 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:39:42 crc kubenswrapper[4690]: E1203 13:39:42.315619 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:39:42 crc kubenswrapper[4690]: I1203 13:39:42.975459 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" event={"ID":"d4625caf-d841-4682-ae69-e7d2d90ef220","Type":"ContainerStarted","Data":"a50103dbda9e4d0ab8da29d4ab3adee7f1473c4bd43522dba7a0ef0536708b52"} Dec 03 13:39:43 crc kubenswrapper[4690]: I1203 13:39:43.041074 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" podStartSLOduration=1.5988678539999999 podStartE2EDuration="2.04104455s" podCreationTimestamp="2025-12-03 13:39:41 +0000 UTC" firstStartedPulling="2025-12-03 13:39:41.935202784 +0000 UTC m=+1827.916123217" lastFinishedPulling="2025-12-03 13:39:42.37737947 +0000 UTC m=+1828.358299913" observedRunningTime="2025-12-03 13:39:43.031160685 +0000 UTC m=+1829.012081128" watchObservedRunningTime="2025-12-03 13:39:43.04104455 +0000 UTC m=+1829.021964983" Dec 03 13:39:46 crc kubenswrapper[4690]: I1203 13:39:46.017671 4690 generic.go:334] "Generic (PLEG): container finished" podID="d4625caf-d841-4682-ae69-e7d2d90ef220" containerID="a50103dbda9e4d0ab8da29d4ab3adee7f1473c4bd43522dba7a0ef0536708b52" exitCode=0 Dec 03 13:39:46 crc kubenswrapper[4690]: I1203 13:39:46.017756 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" event={"ID":"d4625caf-d841-4682-ae69-e7d2d90ef220","Type":"ContainerDied","Data":"a50103dbda9e4d0ab8da29d4ab3adee7f1473c4bd43522dba7a0ef0536708b52"} Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.510744 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.517978 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-ssh-key\") pod \"d4625caf-d841-4682-ae69-e7d2d90ef220\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.518260 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-inventory\") pod \"d4625caf-d841-4682-ae69-e7d2d90ef220\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.518318 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c92p\" (UniqueName: \"kubernetes.io/projected/d4625caf-d841-4682-ae69-e7d2d90ef220-kube-api-access-4c92p\") pod \"d4625caf-d841-4682-ae69-e7d2d90ef220\" (UID: \"d4625caf-d841-4682-ae69-e7d2d90ef220\") " Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.527399 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4625caf-d841-4682-ae69-e7d2d90ef220-kube-api-access-4c92p" (OuterVolumeSpecName: "kube-api-access-4c92p") pod "d4625caf-d841-4682-ae69-e7d2d90ef220" (UID: "d4625caf-d841-4682-ae69-e7d2d90ef220"). InnerVolumeSpecName "kube-api-access-4c92p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.560860 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-inventory" (OuterVolumeSpecName: "inventory") pod "d4625caf-d841-4682-ae69-e7d2d90ef220" (UID: "d4625caf-d841-4682-ae69-e7d2d90ef220"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.562843 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d4625caf-d841-4682-ae69-e7d2d90ef220" (UID: "d4625caf-d841-4682-ae69-e7d2d90ef220"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.623770 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.623827 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c92p\" (UniqueName: \"kubernetes.io/projected/d4625caf-d841-4682-ae69-e7d2d90ef220-kube-api-access-4c92p\") on node \"crc\" DevicePath \"\"" Dec 03 13:39:47 crc kubenswrapper[4690]: I1203 13:39:47.623838 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4625caf-d841-4682-ae69-e7d2d90ef220-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.039013 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" event={"ID":"d4625caf-d841-4682-ae69-e7d2d90ef220","Type":"ContainerDied","Data":"8318867fd1e4d6b5fa5469224284d9c4a5d377a35529d8efa614f08720eaacd1"} Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.039056 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8318867fd1e4d6b5fa5469224284d9c4a5d377a35529d8efa614f08720eaacd1" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.039128 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f924x" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.116030 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl"] Dec 03 13:39:48 crc kubenswrapper[4690]: E1203 13:39:48.116522 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4625caf-d841-4682-ae69-e7d2d90ef220" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.116542 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4625caf-d841-4682-ae69-e7d2d90ef220" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.116757 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4625caf-d841-4682-ae69-e7d2d90ef220" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.117728 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.121516 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.121764 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.121785 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.122102 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.132745 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s56j6\" (UniqueName: \"kubernetes.io/projected/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-kube-api-access-s56j6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.132829 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.133149 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.133546 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.144905 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl"] Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.236017 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.236412 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.236455 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s56j6\" (UniqueName: \"kubernetes.io/projected/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-kube-api-access-s56j6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.236499 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.241593 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.242049 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.243681 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.252347 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s56j6\" (UniqueName: \"kubernetes.io/projected/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-kube-api-access-s56j6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.450060 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:39:48 crc kubenswrapper[4690]: I1203 13:39:48.994976 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl"] Dec 03 13:39:49 crc kubenswrapper[4690]: I1203 13:39:49.051333 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" event={"ID":"ecded0ed-1510-4ac8-80d0-dd7b06c8446e","Type":"ContainerStarted","Data":"a2ea2e7d41f8dac5e9234c38ba9794683f3d0f79179666f61fa7a97546b3cac7"} Dec 03 13:39:50 crc kubenswrapper[4690]: I1203 13:39:50.070598 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" event={"ID":"ecded0ed-1510-4ac8-80d0-dd7b06c8446e","Type":"ContainerStarted","Data":"bc70f20107e4d9dec685e191b482bc097d5fe0b835cc53b3b85284bfba231f34"} Dec 03 13:39:50 crc kubenswrapper[4690]: I1203 13:39:50.092535 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" podStartSLOduration=1.592256232 podStartE2EDuration="2.092515607s" podCreationTimestamp="2025-12-03 13:39:48 +0000 UTC" firstStartedPulling="2025-12-03 13:39:48.999232325 +0000 UTC m=+1834.980152758" lastFinishedPulling="2025-12-03 13:39:49.4994917 +0000 UTC m=+1835.480412133" observedRunningTime="2025-12-03 13:39:50.08563284 +0000 UTC m=+1836.066553273" watchObservedRunningTime="2025-12-03 13:39:50.092515607 +0000 UTC m=+1836.073436040" Dec 03 13:39:55 crc kubenswrapper[4690]: I1203 13:39:55.316800 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:39:55 crc kubenswrapper[4690]: E1203 13:39:55.317776 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:40:08 crc kubenswrapper[4690]: I1203 13:40:08.315473 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:40:08 crc kubenswrapper[4690]: E1203 13:40:08.318017 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:40:21 crc kubenswrapper[4690]: I1203 13:40:21.315537 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:40:21 crc kubenswrapper[4690]: E1203 13:40:21.316639 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:40:32 crc kubenswrapper[4690]: I1203 13:40:32.315499 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:40:32 crc kubenswrapper[4690]: E1203 13:40:32.316908 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:40:43 crc kubenswrapper[4690]: I1203 13:40:43.314980 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:40:43 crc kubenswrapper[4690]: E1203 13:40:43.316036 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.054220 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-rf676"] Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.065475 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-rf676"] Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.076170 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5bb56"] Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.085999 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-wn8pw"] Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.095980 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-wn8pw"] Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.105540 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5bb56"] Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.334434 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca950cf-73a9-47ae-b07c-b8806057e24b" path="/var/lib/kubelet/pods/3ca950cf-73a9-47ae-b07c-b8806057e24b/volumes" Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.335019 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bb598ea-0a2a-49b6-a989-533ab58cd524" path="/var/lib/kubelet/pods/6bb598ea-0a2a-49b6-a989-533ab58cd524/volumes" Dec 03 13:40:56 crc kubenswrapper[4690]: I1203 13:40:56.335559 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c88c34d-c55a-4456-9f8b-daf83b54896a" path="/var/lib/kubelet/pods/9c88c34d-c55a-4456-9f8b-daf83b54896a/volumes" Dec 03 13:40:57 crc kubenswrapper[4690]: I1203 13:40:57.027232 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a217-account-create-update-9pg4q"] Dec 03 13:40:57 crc kubenswrapper[4690]: I1203 13:40:57.046149 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a217-account-create-update-9pg4q"] Dec 03 13:40:58 crc kubenswrapper[4690]: I1203 13:40:58.037994 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-deae-account-create-update-49dp2"] Dec 03 13:40:58 crc kubenswrapper[4690]: I1203 13:40:58.052775 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-deae-account-create-update-49dp2"] Dec 03 13:40:58 crc kubenswrapper[4690]: I1203 13:40:58.064229 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8784-account-create-update-nc4fz"] Dec 03 13:40:58 crc kubenswrapper[4690]: I1203 13:40:58.078314 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8784-account-create-update-nc4fz"] Dec 03 13:40:58 crc kubenswrapper[4690]: I1203 13:40:58.315849 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:40:58 crc kubenswrapper[4690]: E1203 13:40:58.316422 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:40:58 crc kubenswrapper[4690]: I1203 13:40:58.327929 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54bfb9d7-bab2-4541-abd5-96f850bffdde" path="/var/lib/kubelet/pods/54bfb9d7-bab2-4541-abd5-96f850bffdde/volumes" Dec 03 13:40:58 crc kubenswrapper[4690]: I1203 13:40:58.328711 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cf94540-efe2-4e03-80ff-05dbef2b7566" path="/var/lib/kubelet/pods/9cf94540-efe2-4e03-80ff-05dbef2b7566/volumes" Dec 03 13:40:58 crc kubenswrapper[4690]: I1203 13:40:58.329289 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aac7e979-c496-4a3a-8cf0-d3a34d659b85" path="/var/lib/kubelet/pods/aac7e979-c496-4a3a-8cf0-d3a34d659b85/volumes" Dec 03 13:41:09 crc kubenswrapper[4690]: I1203 13:41:09.316083 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:41:09 crc kubenswrapper[4690]: E1203 13:41:09.317318 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:41:22 crc kubenswrapper[4690]: I1203 13:41:22.047233 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-hbltr"] Dec 03 13:41:22 crc kubenswrapper[4690]: I1203 13:41:22.057469 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-hbltr"] Dec 03 13:41:22 crc kubenswrapper[4690]: I1203 13:41:22.331514 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a58de3c0-aba8-49ce-a59c-67732420e08f" path="/var/lib/kubelet/pods/a58de3c0-aba8-49ce-a59c-67732420e08f/volumes" Dec 03 13:41:24 crc kubenswrapper[4690]: I1203 13:41:24.325900 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:41:24 crc kubenswrapper[4690]: E1203 13:41:24.327732 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.043367 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-dab4-account-create-update-445qb"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.057153 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c6af-account-create-update-bzzjg"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.071995 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-59mk8"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.082522 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-gx6kj"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.095179 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-dab4-account-create-update-445qb"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.105161 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c6af-account-create-update-bzzjg"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.114486 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-gx6kj"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.123894 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-4898-account-create-update-66xl7"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.132087 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-59mk8"] Dec 03 13:41:29 crc kubenswrapper[4690]: I1203 13:41:29.139509 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-4898-account-create-update-66xl7"] Dec 03 13:41:30 crc kubenswrapper[4690]: I1203 13:41:30.334582 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c3e3956-3bd9-4693-9fd1-5040c9f90950" path="/var/lib/kubelet/pods/4c3e3956-3bd9-4693-9fd1-5040c9f90950/volumes" Dec 03 13:41:30 crc kubenswrapper[4690]: I1203 13:41:30.335303 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="769615ff-2aa5-41b8-812b-a95a27f72727" path="/var/lib/kubelet/pods/769615ff-2aa5-41b8-812b-a95a27f72727/volumes" Dec 03 13:41:30 crc kubenswrapper[4690]: I1203 13:41:30.335942 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8441b637-1109-48c4-8bd5-efbcd4a34a56" path="/var/lib/kubelet/pods/8441b637-1109-48c4-8bd5-efbcd4a34a56/volumes" Dec 03 13:41:30 crc kubenswrapper[4690]: I1203 13:41:30.336586 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad027298-b2ff-4145-9a45-2dcbf5b5f228" path="/var/lib/kubelet/pods/ad027298-b2ff-4145-9a45-2dcbf5b5f228/volumes" Dec 03 13:41:30 crc kubenswrapper[4690]: I1203 13:41:30.337957 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b204d5-9b12-409c-ac51-6a14dc45debe" path="/var/lib/kubelet/pods/d7b204d5-9b12-409c-ac51-6a14dc45debe/volumes" Dec 03 13:41:36 crc kubenswrapper[4690]: I1203 13:41:36.315683 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:41:36 crc kubenswrapper[4690]: E1203 13:41:36.317212 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:41:39 crc kubenswrapper[4690]: I1203 13:41:39.863768 4690 scope.go:117] "RemoveContainer" containerID="e5ebb4907d1025c203c301da08555282a651966549e0b7275e8be1784bdf9d89" Dec 03 13:41:39 crc kubenswrapper[4690]: I1203 13:41:39.888967 4690 scope.go:117] "RemoveContainer" containerID="630a2298b07ad935706b0bfcb75059cf6a14df30bf6a0924854fdcd10094049f" Dec 03 13:41:39 crc kubenswrapper[4690]: I1203 13:41:39.916830 4690 scope.go:117] "RemoveContainer" containerID="cab2422914baf7975e011762ba51c134d42cd25c400e6a0a141d69b1d1845292" Dec 03 13:41:39 crc kubenswrapper[4690]: I1203 13:41:39.969257 4690 scope.go:117] "RemoveContainer" containerID="33d38665059fcd2b25d5b957f41faa322872860d1668172ae577b70954264b5a" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.019551 4690 scope.go:117] "RemoveContainer" containerID="c1042ec07c23e9f063277430ca0b7cfd0d67cdcd5665c98a64703096e959dd03" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.077003 4690 scope.go:117] "RemoveContainer" containerID="ec874ce6809ef9799dbfd91855c47d4be6871ade2e1f1816dc7f3a10f915a9a0" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.128925 4690 scope.go:117] "RemoveContainer" containerID="99a0b9129c7bb324033787adefc3f16a12ea5f107eaa7cb0d938bce154d446ef" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.171833 4690 scope.go:117] "RemoveContainer" containerID="afe960a664ab75914e0e7966e9fce674c927eae62877287a5ba6cb22b82cb458" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.193798 4690 scope.go:117] "RemoveContainer" containerID="a1f5b9092d0f7088b1f4e64d7af42c003768d24cba564cbea668ffb8a107dd98" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.226584 4690 scope.go:117] "RemoveContainer" containerID="a20e14f317c449e34abb455815a00507875b98e9afbc46c3683065de31818ec7" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.247664 4690 scope.go:117] "RemoveContainer" containerID="d57dfb990b6a701faa8a781c512958741ff80692d4ce3558add0eaff454a2a27" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.265161 4690 scope.go:117] "RemoveContainer" containerID="95172ea3e37dbdc4f54830ad04cd06abe1f20fd9e0eb60fd78fb9765d7a8a47a" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.282025 4690 scope.go:117] "RemoveContainer" containerID="45cd8e316e6636de21df426cb31d01e567b19ad8140be8156f77dfd45ea4b35c" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.302395 4690 scope.go:117] "RemoveContainer" containerID="d5a560113cb0c1ddb073b28d1c0ebdceb9d951d0f23e024af91b5054e49a45d8" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.332485 4690 scope.go:117] "RemoveContainer" containerID="a7afb4538e398dd6c406596836c6ce88468bd7d7a96425af59939745d0c9e93d" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.352471 4690 scope.go:117] "RemoveContainer" containerID="44255d27ab5f10c96b4a3fc74811b968b3b155c4c7baa87b82878e14b3c08cd3" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.381011 4690 scope.go:117] "RemoveContainer" containerID="3a786c0c07d6ddec354cc5ecbe3e7e12dc2108dc64543a9211a0ea72586f7d5b" Dec 03 13:41:40 crc kubenswrapper[4690]: I1203 13:41:40.404680 4690 scope.go:117] "RemoveContainer" containerID="b4bf607df358d77e16ffd4970d7128e96cfda86b194a0aeb0d5fea1f0bed72a0" Dec 03 13:41:46 crc kubenswrapper[4690]: I1203 13:41:46.042820 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-cg7p2"] Dec 03 13:41:46 crc kubenswrapper[4690]: I1203 13:41:46.057803 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-cg7p2"] Dec 03 13:41:46 crc kubenswrapper[4690]: I1203 13:41:46.327932 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="153bc447-b9f4-4746-9869-17088a3b71ae" path="/var/lib/kubelet/pods/153bc447-b9f4-4746-9869-17088a3b71ae/volumes" Dec 03 13:41:48 crc kubenswrapper[4690]: I1203 13:41:48.315336 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:41:48 crc kubenswrapper[4690]: E1203 13:41:48.316100 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:42:01 crc kubenswrapper[4690]: I1203 13:42:01.315036 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:42:01 crc kubenswrapper[4690]: E1203 13:42:01.316116 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:42:12 crc kubenswrapper[4690]: I1203 13:42:12.037275 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-fq5zp"] Dec 03 13:42:12 crc kubenswrapper[4690]: I1203 13:42:12.054178 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-fq5zp"] Dec 03 13:42:12 crc kubenswrapper[4690]: I1203 13:42:12.315509 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:42:12 crc kubenswrapper[4690]: E1203 13:42:12.315814 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:42:12 crc kubenswrapper[4690]: I1203 13:42:12.329753 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1a38ea3-0bc0-4533-bbec-de05b19433c7" path="/var/lib/kubelet/pods/f1a38ea3-0bc0-4533-bbec-de05b19433c7/volumes" Dec 03 13:42:21 crc kubenswrapper[4690]: I1203 13:42:21.055491 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9srzs"] Dec 03 13:42:21 crc kubenswrapper[4690]: I1203 13:42:21.064692 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9srzs"] Dec 03 13:42:22 crc kubenswrapper[4690]: I1203 13:42:22.329002 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e754a7fd-8dd6-4ac4-b197-9c26c946d540" path="/var/lib/kubelet/pods/e754a7fd-8dd6-4ac4-b197-9c26c946d540/volumes" Dec 03 13:42:27 crc kubenswrapper[4690]: I1203 13:42:27.316325 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:42:27 crc kubenswrapper[4690]: I1203 13:42:27.797533 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"4587129f3424bae3d234d695d33b32c949d4b82d91b7d56a0e8ebc65700b9dec"} Dec 03 13:42:40 crc kubenswrapper[4690]: I1203 13:42:40.659012 4690 scope.go:117] "RemoveContainer" containerID="e686b923d12a143f1dff83d5a8f41d5c3566f5144b2f0b912ca84f5d26ac33ef" Dec 03 13:42:40 crc kubenswrapper[4690]: I1203 13:42:40.705075 4690 scope.go:117] "RemoveContainer" containerID="da2dee4909c61fe2e7fc186fa657a3df3adf97f4c8a6b05fba3b41a855c10360" Dec 03 13:42:40 crc kubenswrapper[4690]: I1203 13:42:40.810421 4690 scope.go:117] "RemoveContainer" containerID="5df4f0ea5dcf8aa4c3d6c016fb6a7094f3e4b79be97a3e8e862d4a02fff5e276" Dec 03 13:42:46 crc kubenswrapper[4690]: I1203 13:42:46.051334 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-2vcg6"] Dec 03 13:42:46 crc kubenswrapper[4690]: I1203 13:42:46.068559 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-2vcg6"] Dec 03 13:42:46 crc kubenswrapper[4690]: I1203 13:42:46.329093 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29bf3e45-7939-4b48-923b-d0f28bf5c046" path="/var/lib/kubelet/pods/29bf3e45-7939-4b48-923b-d0f28bf5c046/volumes" Dec 03 13:42:50 crc kubenswrapper[4690]: I1203 13:42:50.040999 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ndpp9"] Dec 03 13:42:50 crc kubenswrapper[4690]: I1203 13:42:50.049525 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ndpp9"] Dec 03 13:42:50 crc kubenswrapper[4690]: I1203 13:42:50.332071 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62230b01-b33e-4213-80c3-ecf6d5f995b1" path="/var/lib/kubelet/pods/62230b01-b33e-4213-80c3-ecf6d5f995b1/volumes" Dec 03 13:43:16 crc kubenswrapper[4690]: I1203 13:43:16.050685 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-2rlzz"] Dec 03 13:43:16 crc kubenswrapper[4690]: I1203 13:43:16.064461 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-2rlzz"] Dec 03 13:43:16 crc kubenswrapper[4690]: I1203 13:43:16.330670 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88d2e02a-af82-49f0-858a-9e9fa3dc4985" path="/var/lib/kubelet/pods/88d2e02a-af82-49f0-858a-9e9fa3dc4985/volumes" Dec 03 13:43:39 crc kubenswrapper[4690]: I1203 13:43:39.032600 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pzvps"] Dec 03 13:43:39 crc kubenswrapper[4690]: I1203 13:43:39.043202 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-8k487"] Dec 03 13:43:39 crc kubenswrapper[4690]: I1203 13:43:39.052429 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-8k487"] Dec 03 13:43:39 crc kubenswrapper[4690]: I1203 13:43:39.060416 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pzvps"] Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.034112 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5b13-account-create-update-w7g9f"] Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.044397 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-rxz2w"] Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.054144 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5b13-account-create-update-w7g9f"] Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.062679 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-rxz2w"] Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.327573 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2478ef86-8c11-4d36-88b1-eea0a3b2310a" path="/var/lib/kubelet/pods/2478ef86-8c11-4d36-88b1-eea0a3b2310a/volumes" Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.328677 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8137b9f0-e7c6-4134-ae61-5599982fa141" path="/var/lib/kubelet/pods/8137b9f0-e7c6-4134-ae61-5599982fa141/volumes" Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.329604 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad6e77ea-b90d-4def-8ab1-d761fc41ebe3" path="/var/lib/kubelet/pods/ad6e77ea-b90d-4def-8ab1-d761fc41ebe3/volumes" Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.330560 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbddf981-d575-4517-b0e6-f494b9a92fe6" path="/var/lib/kubelet/pods/dbddf981-d575-4517-b0e6-f494b9a92fe6/volumes" Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.921665 4690 scope.go:117] "RemoveContainer" containerID="84d532dd714b5dbf5bc6ee00db6886ddf3ebe64dfec1465c5d61f5ac86193c30" Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.955541 4690 scope.go:117] "RemoveContainer" containerID="6a91b52845253aef44e053f1c33a5981cee01a273677f375c5fc8ff8e7539d3d" Dec 03 13:43:40 crc kubenswrapper[4690]: I1203 13:43:40.994016 4690 scope.go:117] "RemoveContainer" containerID="79843863f21b26dc627742ae974accc39c1d70ce7eb185efa682ff1e93cd2def" Dec 03 13:43:41 crc kubenswrapper[4690]: I1203 13:43:41.029931 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8cf5-account-create-update-2m8zh"] Dec 03 13:43:41 crc kubenswrapper[4690]: I1203 13:43:41.059330 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8cf5-account-create-update-2m8zh"] Dec 03 13:43:41 crc kubenswrapper[4690]: I1203 13:43:41.066194 4690 scope.go:117] "RemoveContainer" containerID="2a92c73f8d3c14ce13f0f46cd1e56c15be1d86522956396c59edcce9221675b4" Dec 03 13:43:41 crc kubenswrapper[4690]: I1203 13:43:41.106595 4690 scope.go:117] "RemoveContainer" containerID="c052f9f5ff262734453b7188da832580e8edd7fd7095639527e4e1ba04504bed" Dec 03 13:43:41 crc kubenswrapper[4690]: I1203 13:43:41.186689 4690 scope.go:117] "RemoveContainer" containerID="be3917a1f502330f72af2766300a0420d1fbdd574129993d9051d6ac51c7b363" Dec 03 13:43:41 crc kubenswrapper[4690]: I1203 13:43:41.233225 4690 scope.go:117] "RemoveContainer" containerID="6e249d4455c2ded66f572068a6a54d1a5c7e0e1effa4e253c7fd6e290b7d49c2" Dec 03 13:43:42 crc kubenswrapper[4690]: I1203 13:43:42.329301 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bf7fd9d-669f-4d15-a432-8b796d5f1de5" path="/var/lib/kubelet/pods/5bf7fd9d-669f-4d15-a432-8b796d5f1de5/volumes" Dec 03 13:43:43 crc kubenswrapper[4690]: I1203 13:43:43.031743 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-de32-account-create-update-d5v2x"] Dec 03 13:43:43 crc kubenswrapper[4690]: I1203 13:43:43.042323 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-de32-account-create-update-d5v2x"] Dec 03 13:43:44 crc kubenswrapper[4690]: I1203 13:43:44.327580 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8" path="/var/lib/kubelet/pods/16a6fee3-66a8-42cd-8c9f-6bfcf592c8f8/volumes" Dec 03 13:44:02 crc kubenswrapper[4690]: I1203 13:44:02.835075 4690 generic.go:334] "Generic (PLEG): container finished" podID="ecded0ed-1510-4ac8-80d0-dd7b06c8446e" containerID="bc70f20107e4d9dec685e191b482bc097d5fe0b835cc53b3b85284bfba231f34" exitCode=0 Dec 03 13:44:02 crc kubenswrapper[4690]: I1203 13:44:02.835123 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" event={"ID":"ecded0ed-1510-4ac8-80d0-dd7b06c8446e","Type":"ContainerDied","Data":"bc70f20107e4d9dec685e191b482bc097d5fe0b835cc53b3b85284bfba231f34"} Dec 03 13:44:04 crc kubenswrapper[4690]: I1203 13:44:04.348314 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.466273 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-ssh-key\") pod \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.466513 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-inventory\") pod \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.466572 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s56j6\" (UniqueName: \"kubernetes.io/projected/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-kube-api-access-s56j6\") pod \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.466600 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-bootstrap-combined-ca-bundle\") pod \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\" (UID: \"ecded0ed-1510-4ac8-80d0-dd7b06c8446e\") " Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.534605 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-kube-api-access-s56j6" (OuterVolumeSpecName: "kube-api-access-s56j6") pod "ecded0ed-1510-4ac8-80d0-dd7b06c8446e" (UID: "ecded0ed-1510-4ac8-80d0-dd7b06c8446e"). InnerVolumeSpecName "kube-api-access-s56j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.534698 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ecded0ed-1510-4ac8-80d0-dd7b06c8446e" (UID: "ecded0ed-1510-4ac8-80d0-dd7b06c8446e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.567026 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ecded0ed-1510-4ac8-80d0-dd7b06c8446e" (UID: "ecded0ed-1510-4ac8-80d0-dd7b06c8446e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.568656 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.568682 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s56j6\" (UniqueName: \"kubernetes.io/projected/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-kube-api-access-s56j6\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.568697 4690 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.575060 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-inventory" (OuterVolumeSpecName: "inventory") pod "ecded0ed-1510-4ac8-80d0-dd7b06c8446e" (UID: "ecded0ed-1510-4ac8-80d0-dd7b06c8446e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.670711 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecded0ed-1510-4ac8-80d0-dd7b06c8446e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.857343 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" event={"ID":"ecded0ed-1510-4ac8-80d0-dd7b06c8446e","Type":"ContainerDied","Data":"a2ea2e7d41f8dac5e9234c38ba9794683f3d0f79179666f61fa7a97546b3cac7"} Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.857391 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2ea2e7d41f8dac5e9234c38ba9794683f3d0f79179666f61fa7a97546b3cac7" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.857483 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.949535 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh"] Dec 03 13:44:05 crc kubenswrapper[4690]: E1203 13:44:04.949994 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecded0ed-1510-4ac8-80d0-dd7b06c8446e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.950008 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecded0ed-1510-4ac8-80d0-dd7b06c8446e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.950253 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecded0ed-1510-4ac8-80d0-dd7b06c8446e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.951163 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.953573 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.954404 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.958790 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.964193 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh"] Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:04.967087 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.078529 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.078573 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.078776 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhzcs\" (UniqueName: \"kubernetes.io/projected/2f30c127-ad2e-4154-a344-61a21b57b3d0-kube-api-access-hhzcs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.180943 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhzcs\" (UniqueName: \"kubernetes.io/projected/2f30c127-ad2e-4154-a344-61a21b57b3d0-kube-api-access-hhzcs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.181038 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.181066 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.185730 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.185759 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.199230 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhzcs\" (UniqueName: \"kubernetes.io/projected/2f30c127-ad2e-4154-a344-61a21b57b3d0-kube-api-access-hhzcs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pksbh\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.272447 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.838999 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh"] Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.849330 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:44:05 crc kubenswrapper[4690]: I1203 13:44:05.868853 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" event={"ID":"2f30c127-ad2e-4154-a344-61a21b57b3d0","Type":"ContainerStarted","Data":"4bc448cd2bb9c771c0bd2649963ce5a490f4ae72a4af2d72e66c3ef8706df027"} Dec 03 13:44:06 crc kubenswrapper[4690]: I1203 13:44:06.879121 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" event={"ID":"2f30c127-ad2e-4154-a344-61a21b57b3d0","Type":"ContainerStarted","Data":"6e0909f73af76e8ab0baf89a6043aabb38b5a6b628a96817732ff9399c435d3d"} Dec 03 13:44:06 crc kubenswrapper[4690]: I1203 13:44:06.894383 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" podStartSLOduration=2.311760881 podStartE2EDuration="2.894363857s" podCreationTimestamp="2025-12-03 13:44:04 +0000 UTC" firstStartedPulling="2025-12-03 13:44:05.849057169 +0000 UTC m=+2091.829977602" lastFinishedPulling="2025-12-03 13:44:06.431660135 +0000 UTC m=+2092.412580578" observedRunningTime="2025-12-03 13:44:06.892768656 +0000 UTC m=+2092.873689109" watchObservedRunningTime="2025-12-03 13:44:06.894363857 +0000 UTC m=+2092.875284290" Dec 03 13:44:13 crc kubenswrapper[4690]: I1203 13:44:13.042296 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2xpmv"] Dec 03 13:44:13 crc kubenswrapper[4690]: I1203 13:44:13.054545 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2xpmv"] Dec 03 13:44:14 crc kubenswrapper[4690]: I1203 13:44:14.331947 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca321dd-bd9d-4039-aa80-eb386d63dcb4" path="/var/lib/kubelet/pods/3ca321dd-bd9d-4039-aa80-eb386d63dcb4/volumes" Dec 03 13:44:41 crc kubenswrapper[4690]: I1203 13:44:41.378271 4690 scope.go:117] "RemoveContainer" containerID="26f89748c95d5f3a2f242b46d57fe3a6c937a90a0140641b5f163e93f90407d7" Dec 03 13:44:41 crc kubenswrapper[4690]: I1203 13:44:41.407903 4690 scope.go:117] "RemoveContainer" containerID="f82022919bfa95568653fc71076b34c6a96a848f2fd2ebd9e87bb884d122a26b" Dec 03 13:44:41 crc kubenswrapper[4690]: I1203 13:44:41.468094 4690 scope.go:117] "RemoveContainer" containerID="eb7088271fec7391af00bc7bb58cc71b4f9de4efd677b58db0288059fc39f036" Dec 03 13:44:46 crc kubenswrapper[4690]: I1203 13:44:46.823637 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:44:46 crc kubenswrapper[4690]: I1203 13:44:46.824419 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.158642 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr"] Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.162126 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.169097 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.170062 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.172772 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr"] Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.343744 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5717c7dc-0f59-490f-8b60-33b8275899b6-config-volume\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.344167 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlkwp\" (UniqueName: \"kubernetes.io/projected/5717c7dc-0f59-490f-8b60-33b8275899b6-kube-api-access-hlkwp\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.344325 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5717c7dc-0f59-490f-8b60-33b8275899b6-secret-volume\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.446483 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlkwp\" (UniqueName: \"kubernetes.io/projected/5717c7dc-0f59-490f-8b60-33b8275899b6-kube-api-access-hlkwp\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.446552 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5717c7dc-0f59-490f-8b60-33b8275899b6-secret-volume\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.446710 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5717c7dc-0f59-490f-8b60-33b8275899b6-config-volume\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.447500 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5717c7dc-0f59-490f-8b60-33b8275899b6-config-volume\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.456513 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5717c7dc-0f59-490f-8b60-33b8275899b6-secret-volume\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.475721 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlkwp\" (UniqueName: \"kubernetes.io/projected/5717c7dc-0f59-490f-8b60-33b8275899b6-kube-api-access-hlkwp\") pod \"collect-profiles-29412825-n2wnr\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.489727 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:00 crc kubenswrapper[4690]: I1203 13:45:00.971300 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr"] Dec 03 13:45:01 crc kubenswrapper[4690]: I1203 13:45:01.424929 4690 generic.go:334] "Generic (PLEG): container finished" podID="5717c7dc-0f59-490f-8b60-33b8275899b6" containerID="abaf56525d7cdf3965da80f820ec2ca8381781de1172d5ee438d1f5d2ce161e0" exitCode=0 Dec 03 13:45:01 crc kubenswrapper[4690]: I1203 13:45:01.425094 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" event={"ID":"5717c7dc-0f59-490f-8b60-33b8275899b6","Type":"ContainerDied","Data":"abaf56525d7cdf3965da80f820ec2ca8381781de1172d5ee438d1f5d2ce161e0"} Dec 03 13:45:01 crc kubenswrapper[4690]: I1203 13:45:01.427202 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" event={"ID":"5717c7dc-0f59-490f-8b60-33b8275899b6","Type":"ContainerStarted","Data":"f76229576d2c8da6706fd60c4864bd55186a5e20703f873cb8a39909f95b4106"} Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.056038 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.221478 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5717c7dc-0f59-490f-8b60-33b8275899b6-secret-volume\") pod \"5717c7dc-0f59-490f-8b60-33b8275899b6\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.221644 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlkwp\" (UniqueName: \"kubernetes.io/projected/5717c7dc-0f59-490f-8b60-33b8275899b6-kube-api-access-hlkwp\") pod \"5717c7dc-0f59-490f-8b60-33b8275899b6\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.221703 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5717c7dc-0f59-490f-8b60-33b8275899b6-config-volume\") pod \"5717c7dc-0f59-490f-8b60-33b8275899b6\" (UID: \"5717c7dc-0f59-490f-8b60-33b8275899b6\") " Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.222933 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5717c7dc-0f59-490f-8b60-33b8275899b6-config-volume" (OuterVolumeSpecName: "config-volume") pod "5717c7dc-0f59-490f-8b60-33b8275899b6" (UID: "5717c7dc-0f59-490f-8b60-33b8275899b6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.223902 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5717c7dc-0f59-490f-8b60-33b8275899b6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.229194 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5717c7dc-0f59-490f-8b60-33b8275899b6-kube-api-access-hlkwp" (OuterVolumeSpecName: "kube-api-access-hlkwp") pod "5717c7dc-0f59-490f-8b60-33b8275899b6" (UID: "5717c7dc-0f59-490f-8b60-33b8275899b6"). InnerVolumeSpecName "kube-api-access-hlkwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.229410 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5717c7dc-0f59-490f-8b60-33b8275899b6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5717c7dc-0f59-490f-8b60-33b8275899b6" (UID: "5717c7dc-0f59-490f-8b60-33b8275899b6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.325992 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5717c7dc-0f59-490f-8b60-33b8275899b6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.326036 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlkwp\" (UniqueName: \"kubernetes.io/projected/5717c7dc-0f59-490f-8b60-33b8275899b6-kube-api-access-hlkwp\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.456373 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" event={"ID":"5717c7dc-0f59-490f-8b60-33b8275899b6","Type":"ContainerDied","Data":"f76229576d2c8da6706fd60c4864bd55186a5e20703f873cb8a39909f95b4106"} Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.456428 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f76229576d2c8da6706fd60c4864bd55186a5e20703f873cb8a39909f95b4106" Dec 03 13:45:03 crc kubenswrapper[4690]: I1203 13:45:03.456437 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr" Dec 03 13:45:04 crc kubenswrapper[4690]: I1203 13:45:04.152734 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk"] Dec 03 13:45:04 crc kubenswrapper[4690]: I1203 13:45:04.165250 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-726dk"] Dec 03 13:45:04 crc kubenswrapper[4690]: I1203 13:45:04.330817 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23e81a05-bdfd-4a30-82b0-9eded242e91c" path="/var/lib/kubelet/pods/23e81a05-bdfd-4a30-82b0-9eded242e91c/volumes" Dec 03 13:45:12 crc kubenswrapper[4690]: I1203 13:45:12.040006 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hvqqp"] Dec 03 13:45:12 crc kubenswrapper[4690]: I1203 13:45:12.052400 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hvqqp"] Dec 03 13:45:12 crc kubenswrapper[4690]: I1203 13:45:12.329854 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43129653-d1cd-4ced-b120-33be343c3d9d" path="/var/lib/kubelet/pods/43129653-d1cd-4ced-b120-33be343c3d9d/volumes" Dec 03 13:45:16 crc kubenswrapper[4690]: I1203 13:45:16.823922 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:45:16 crc kubenswrapper[4690]: I1203 13:45:16.824660 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:45:29 crc kubenswrapper[4690]: I1203 13:45:29.938919 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwp6x"] Dec 03 13:45:29 crc kubenswrapper[4690]: E1203 13:45:29.940525 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5717c7dc-0f59-490f-8b60-33b8275899b6" containerName="collect-profiles" Dec 03 13:45:29 crc kubenswrapper[4690]: I1203 13:45:29.940544 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="5717c7dc-0f59-490f-8b60-33b8275899b6" containerName="collect-profiles" Dec 03 13:45:29 crc kubenswrapper[4690]: I1203 13:45:29.940798 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="5717c7dc-0f59-490f-8b60-33b8275899b6" containerName="collect-profiles" Dec 03 13:45:29 crc kubenswrapper[4690]: I1203 13:45:29.943379 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:29 crc kubenswrapper[4690]: I1203 13:45:29.979001 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwp6x"] Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.060768 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbg5s\" (UniqueName: \"kubernetes.io/projected/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-kube-api-access-hbg5s\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.060896 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-utilities\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.061537 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-catalog-content\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.165899 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbg5s\" (UniqueName: \"kubernetes.io/projected/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-kube-api-access-hbg5s\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.166013 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-utilities\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.166130 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-catalog-content\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.166642 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-utilities\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.166898 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-catalog-content\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.190378 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbg5s\" (UniqueName: \"kubernetes.io/projected/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-kube-api-access-hbg5s\") pod \"community-operators-nwp6x\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.271521 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:30 crc kubenswrapper[4690]: I1203 13:45:30.810761 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwp6x"] Dec 03 13:45:31 crc kubenswrapper[4690]: I1203 13:45:31.737693 4690 generic.go:334] "Generic (PLEG): container finished" podID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerID="79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954" exitCode=0 Dec 03 13:45:31 crc kubenswrapper[4690]: I1203 13:45:31.737826 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwp6x" event={"ID":"892596a6-cd4f-4121-8d8b-d4ce9f69cb84","Type":"ContainerDied","Data":"79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954"} Dec 03 13:45:31 crc kubenswrapper[4690]: I1203 13:45:31.738186 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwp6x" event={"ID":"892596a6-cd4f-4121-8d8b-d4ce9f69cb84","Type":"ContainerStarted","Data":"500c1d63891e02b180862ecb266295f65142745db65ca677bedea5c70b34bf6b"} Dec 03 13:45:33 crc kubenswrapper[4690]: I1203 13:45:33.761260 4690 generic.go:334] "Generic (PLEG): container finished" podID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerID="6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c" exitCode=0 Dec 03 13:45:33 crc kubenswrapper[4690]: I1203 13:45:33.761329 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwp6x" event={"ID":"892596a6-cd4f-4121-8d8b-d4ce9f69cb84","Type":"ContainerDied","Data":"6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c"} Dec 03 13:45:34 crc kubenswrapper[4690]: I1203 13:45:34.775667 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwp6x" event={"ID":"892596a6-cd4f-4121-8d8b-d4ce9f69cb84","Type":"ContainerStarted","Data":"7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970"} Dec 03 13:45:34 crc kubenswrapper[4690]: I1203 13:45:34.802306 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwp6x" podStartSLOduration=3.269050952 podStartE2EDuration="5.802282685s" podCreationTimestamp="2025-12-03 13:45:29 +0000 UTC" firstStartedPulling="2025-12-03 13:45:31.739892147 +0000 UTC m=+2177.720812570" lastFinishedPulling="2025-12-03 13:45:34.27312386 +0000 UTC m=+2180.254044303" observedRunningTime="2025-12-03 13:45:34.797197576 +0000 UTC m=+2180.778118019" watchObservedRunningTime="2025-12-03 13:45:34.802282685 +0000 UTC m=+2180.783203118" Dec 03 13:45:37 crc kubenswrapper[4690]: I1203 13:45:37.078674 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-2smlf"] Dec 03 13:45:37 crc kubenswrapper[4690]: I1203 13:45:37.099317 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-2smlf"] Dec 03 13:45:38 crc kubenswrapper[4690]: I1203 13:45:38.333321 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4ae925d-b15a-411f-b2b8-64ddbbbdb097" path="/var/lib/kubelet/pods/d4ae925d-b15a-411f-b2b8-64ddbbbdb097/volumes" Dec 03 13:45:40 crc kubenswrapper[4690]: I1203 13:45:40.034268 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jgxms"] Dec 03 13:45:40 crc kubenswrapper[4690]: I1203 13:45:40.042765 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jgxms"] Dec 03 13:45:40 crc kubenswrapper[4690]: I1203 13:45:40.272484 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:40 crc kubenswrapper[4690]: I1203 13:45:40.272929 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:40 crc kubenswrapper[4690]: I1203 13:45:40.334291 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a65eb6-765c-4360-bed0-c06558a1aad6" path="/var/lib/kubelet/pods/76a65eb6-765c-4360-bed0-c06558a1aad6/volumes" Dec 03 13:45:40 crc kubenswrapper[4690]: I1203 13:45:40.335646 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:40 crc kubenswrapper[4690]: I1203 13:45:40.915860 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:40 crc kubenswrapper[4690]: I1203 13:45:40.973105 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwp6x"] Dec 03 13:45:41 crc kubenswrapper[4690]: I1203 13:45:41.591580 4690 scope.go:117] "RemoveContainer" containerID="b6c576df6574c4663931f9eec991081ae18b6f741b1655bb6916370a999de9bd" Dec 03 13:45:41 crc kubenswrapper[4690]: I1203 13:45:41.658091 4690 scope.go:117] "RemoveContainer" containerID="41fa04cdec6857177bb3fc18b10b5e4e11a9beade25ddd077b1f275dc2a49aa8" Dec 03 13:45:41 crc kubenswrapper[4690]: I1203 13:45:41.705942 4690 scope.go:117] "RemoveContainer" containerID="83d3b49c42eccb6643a650c91780e185a7df0744daad78fb2761714a83574e44" Dec 03 13:45:41 crc kubenswrapper[4690]: I1203 13:45:41.773691 4690 scope.go:117] "RemoveContainer" containerID="c78ade4be78d8a2ea8e784f5b1379b4e7bca6e1f4b13b0601e47348a4e41c8e3" Dec 03 13:45:42 crc kubenswrapper[4690]: I1203 13:45:42.908622 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwp6x" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerName="registry-server" containerID="cri-o://7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970" gracePeriod=2 Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.466600 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.520571 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-utilities\") pod \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.520816 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-catalog-content\") pod \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.520978 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbg5s\" (UniqueName: \"kubernetes.io/projected/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-kube-api-access-hbg5s\") pod \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\" (UID: \"892596a6-cd4f-4121-8d8b-d4ce9f69cb84\") " Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.521832 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-utilities" (OuterVolumeSpecName: "utilities") pod "892596a6-cd4f-4121-8d8b-d4ce9f69cb84" (UID: "892596a6-cd4f-4121-8d8b-d4ce9f69cb84"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.530516 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-kube-api-access-hbg5s" (OuterVolumeSpecName: "kube-api-access-hbg5s") pod "892596a6-cd4f-4121-8d8b-d4ce9f69cb84" (UID: "892596a6-cd4f-4121-8d8b-d4ce9f69cb84"). InnerVolumeSpecName "kube-api-access-hbg5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.580591 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "892596a6-cd4f-4121-8d8b-d4ce9f69cb84" (UID: "892596a6-cd4f-4121-8d8b-d4ce9f69cb84"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.623715 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbg5s\" (UniqueName: \"kubernetes.io/projected/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-kube-api-access-hbg5s\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.624001 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.624062 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/892596a6-cd4f-4121-8d8b-d4ce9f69cb84-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.919024 4690 generic.go:334] "Generic (PLEG): container finished" podID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerID="7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970" exitCode=0 Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.919363 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwp6x" event={"ID":"892596a6-cd4f-4121-8d8b-d4ce9f69cb84","Type":"ContainerDied","Data":"7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970"} Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.919404 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwp6x" event={"ID":"892596a6-cd4f-4121-8d8b-d4ce9f69cb84","Type":"ContainerDied","Data":"500c1d63891e02b180862ecb266295f65142745db65ca677bedea5c70b34bf6b"} Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.919425 4690 scope.go:117] "RemoveContainer" containerID="7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.919594 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwp6x" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.941574 4690 scope.go:117] "RemoveContainer" containerID="6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c" Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.964986 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwp6x"] Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.974112 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwp6x"] Dec 03 13:45:43 crc kubenswrapper[4690]: I1203 13:45:43.982217 4690 scope.go:117] "RemoveContainer" containerID="79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954" Dec 03 13:45:44 crc kubenswrapper[4690]: I1203 13:45:44.015564 4690 scope.go:117] "RemoveContainer" containerID="7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970" Dec 03 13:45:44 crc kubenswrapper[4690]: E1203 13:45:44.016240 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970\": container with ID starting with 7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970 not found: ID does not exist" containerID="7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970" Dec 03 13:45:44 crc kubenswrapper[4690]: I1203 13:45:44.016295 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970"} err="failed to get container status \"7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970\": rpc error: code = NotFound desc = could not find container \"7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970\": container with ID starting with 7026c2e7c231f528d545b27c3f9c19a29a9e00b39db14489d88e7e9c50f1c970 not found: ID does not exist" Dec 03 13:45:44 crc kubenswrapper[4690]: I1203 13:45:44.016330 4690 scope.go:117] "RemoveContainer" containerID="6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c" Dec 03 13:45:44 crc kubenswrapper[4690]: E1203 13:45:44.017715 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c\": container with ID starting with 6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c not found: ID does not exist" containerID="6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c" Dec 03 13:45:44 crc kubenswrapper[4690]: I1203 13:45:44.017752 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c"} err="failed to get container status \"6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c\": rpc error: code = NotFound desc = could not find container \"6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c\": container with ID starting with 6b4d8a9acd6a2a4061f0cb99692a6a3d4ec759707878acb08a1dfcc27db7154c not found: ID does not exist" Dec 03 13:45:44 crc kubenswrapper[4690]: I1203 13:45:44.017773 4690 scope.go:117] "RemoveContainer" containerID="79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954" Dec 03 13:45:44 crc kubenswrapper[4690]: E1203 13:45:44.018068 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954\": container with ID starting with 79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954 not found: ID does not exist" containerID="79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954" Dec 03 13:45:44 crc kubenswrapper[4690]: I1203 13:45:44.018100 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954"} err="failed to get container status \"79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954\": rpc error: code = NotFound desc = could not find container \"79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954\": container with ID starting with 79626e38b4e3fbb1b1f3eb14a995b30991f318a564affd207feef962beee7954 not found: ID does not exist" Dec 03 13:45:44 crc kubenswrapper[4690]: I1203 13:45:44.326985 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" path="/var/lib/kubelet/pods/892596a6-cd4f-4121-8d8b-d4ce9f69cb84/volumes" Dec 03 13:45:46 crc kubenswrapper[4690]: I1203 13:45:46.823560 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:45:46 crc kubenswrapper[4690]: I1203 13:45:46.824190 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:45:46 crc kubenswrapper[4690]: I1203 13:45:46.824243 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:45:46 crc kubenswrapper[4690]: I1203 13:45:46.825067 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4587129f3424bae3d234d695d33b32c949d4b82d91b7d56a0e8ebc65700b9dec"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:45:46 crc kubenswrapper[4690]: I1203 13:45:46.825126 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://4587129f3424bae3d234d695d33b32c949d4b82d91b7d56a0e8ebc65700b9dec" gracePeriod=600 Dec 03 13:45:46 crc kubenswrapper[4690]: I1203 13:45:46.962400 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="4587129f3424bae3d234d695d33b32c949d4b82d91b7d56a0e8ebc65700b9dec" exitCode=0 Dec 03 13:45:46 crc kubenswrapper[4690]: I1203 13:45:46.962454 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"4587129f3424bae3d234d695d33b32c949d4b82d91b7d56a0e8ebc65700b9dec"} Dec 03 13:45:46 crc kubenswrapper[4690]: I1203 13:45:46.962491 4690 scope.go:117] "RemoveContainer" containerID="340ac6f628025f9c636107c65911d2739ff2c85b73b0ea1cc99228f3e5877bc2" Dec 03 13:45:47 crc kubenswrapper[4690]: I1203 13:45:47.975635 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1"} Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.521516 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mtp8m"] Dec 03 13:46:02 crc kubenswrapper[4690]: E1203 13:46:02.524033 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerName="extract-content" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.524134 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerName="extract-content" Dec 03 13:46:02 crc kubenswrapper[4690]: E1203 13:46:02.524228 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerName="registry-server" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.524306 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerName="registry-server" Dec 03 13:46:02 crc kubenswrapper[4690]: E1203 13:46:02.524389 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerName="extract-utilities" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.524452 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerName="extract-utilities" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.524728 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="892596a6-cd4f-4121-8d8b-d4ce9f69cb84" containerName="registry-server" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.526489 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.533341 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mtp8m"] Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.684250 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm8j7\" (UniqueName: \"kubernetes.io/projected/303e2849-4633-4722-b3b9-9e87cf3c9ed0-kube-api-access-dm8j7\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.684329 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-catalog-content\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.684543 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-utilities\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.787581 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-catalog-content\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.787898 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-utilities\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.788079 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm8j7\" (UniqueName: \"kubernetes.io/projected/303e2849-4633-4722-b3b9-9e87cf3c9ed0-kube-api-access-dm8j7\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.788458 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-utilities\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.788468 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-catalog-content\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.809954 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm8j7\" (UniqueName: \"kubernetes.io/projected/303e2849-4633-4722-b3b9-9e87cf3c9ed0-kube-api-access-dm8j7\") pod \"certified-operators-mtp8m\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:02 crc kubenswrapper[4690]: I1203 13:46:02.856071 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:03 crc kubenswrapper[4690]: W1203 13:46:03.396512 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod303e2849_4633_4722_b3b9_9e87cf3c9ed0.slice/crio-e605bb586165d44fbf9700adbcb17f3266f2326c49d0eb747b4bf5430070878a WatchSource:0}: Error finding container e605bb586165d44fbf9700adbcb17f3266f2326c49d0eb747b4bf5430070878a: Status 404 returned error can't find the container with id e605bb586165d44fbf9700adbcb17f3266f2326c49d0eb747b4bf5430070878a Dec 03 13:46:03 crc kubenswrapper[4690]: I1203 13:46:03.402784 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mtp8m"] Dec 03 13:46:04 crc kubenswrapper[4690]: I1203 13:46:04.139985 4690 generic.go:334] "Generic (PLEG): container finished" podID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerID="847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd" exitCode=0 Dec 03 13:46:04 crc kubenswrapper[4690]: I1203 13:46:04.140080 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtp8m" event={"ID":"303e2849-4633-4722-b3b9-9e87cf3c9ed0","Type":"ContainerDied","Data":"847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd"} Dec 03 13:46:04 crc kubenswrapper[4690]: I1203 13:46:04.140492 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtp8m" event={"ID":"303e2849-4633-4722-b3b9-9e87cf3c9ed0","Type":"ContainerStarted","Data":"e605bb586165d44fbf9700adbcb17f3266f2326c49d0eb747b4bf5430070878a"} Dec 03 13:46:05 crc kubenswrapper[4690]: I1203 13:46:05.162676 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtp8m" event={"ID":"303e2849-4633-4722-b3b9-9e87cf3c9ed0","Type":"ContainerStarted","Data":"ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1"} Dec 03 13:46:06 crc kubenswrapper[4690]: I1203 13:46:06.174950 4690 generic.go:334] "Generic (PLEG): container finished" podID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerID="ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1" exitCode=0 Dec 03 13:46:06 crc kubenswrapper[4690]: I1203 13:46:06.175000 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtp8m" event={"ID":"303e2849-4633-4722-b3b9-9e87cf3c9ed0","Type":"ContainerDied","Data":"ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1"} Dec 03 13:46:08 crc kubenswrapper[4690]: I1203 13:46:08.196799 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtp8m" event={"ID":"303e2849-4633-4722-b3b9-9e87cf3c9ed0","Type":"ContainerStarted","Data":"be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d"} Dec 03 13:46:08 crc kubenswrapper[4690]: I1203 13:46:08.230611 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mtp8m" podStartSLOduration=2.898999229 podStartE2EDuration="6.230588396s" podCreationTimestamp="2025-12-03 13:46:02 +0000 UTC" firstStartedPulling="2025-12-03 13:46:04.144487232 +0000 UTC m=+2210.125407675" lastFinishedPulling="2025-12-03 13:46:07.476076409 +0000 UTC m=+2213.456996842" observedRunningTime="2025-12-03 13:46:08.21646747 +0000 UTC m=+2214.197387903" watchObservedRunningTime="2025-12-03 13:46:08.230588396 +0000 UTC m=+2214.211508829" Dec 03 13:46:12 crc kubenswrapper[4690]: I1203 13:46:12.856773 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:12 crc kubenswrapper[4690]: I1203 13:46:12.857994 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:12 crc kubenswrapper[4690]: I1203 13:46:12.918120 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:13 crc kubenswrapper[4690]: I1203 13:46:13.294321 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:13 crc kubenswrapper[4690]: I1203 13:46:13.354399 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mtp8m"] Dec 03 13:46:15 crc kubenswrapper[4690]: I1203 13:46:15.264452 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mtp8m" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerName="registry-server" containerID="cri-o://be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d" gracePeriod=2 Dec 03 13:46:15 crc kubenswrapper[4690]: I1203 13:46:15.867369 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:15 crc kubenswrapper[4690]: I1203 13:46:15.919592 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm8j7\" (UniqueName: \"kubernetes.io/projected/303e2849-4633-4722-b3b9-9e87cf3c9ed0-kube-api-access-dm8j7\") pod \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " Dec 03 13:46:15 crc kubenswrapper[4690]: I1203 13:46:15.919765 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-catalog-content\") pod \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " Dec 03 13:46:15 crc kubenswrapper[4690]: I1203 13:46:15.919910 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-utilities\") pod \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\" (UID: \"303e2849-4633-4722-b3b9-9e87cf3c9ed0\") " Dec 03 13:46:15 crc kubenswrapper[4690]: I1203 13:46:15.921025 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-utilities" (OuterVolumeSpecName: "utilities") pod "303e2849-4633-4722-b3b9-9e87cf3c9ed0" (UID: "303e2849-4633-4722-b3b9-9e87cf3c9ed0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:46:15 crc kubenswrapper[4690]: I1203 13:46:15.930131 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/303e2849-4633-4722-b3b9-9e87cf3c9ed0-kube-api-access-dm8j7" (OuterVolumeSpecName: "kube-api-access-dm8j7") pod "303e2849-4633-4722-b3b9-9e87cf3c9ed0" (UID: "303e2849-4633-4722-b3b9-9e87cf3c9ed0"). InnerVolumeSpecName "kube-api-access-dm8j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:46:15 crc kubenswrapper[4690]: I1203 13:46:15.969903 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "303e2849-4633-4722-b3b9-9e87cf3c9ed0" (UID: "303e2849-4633-4722-b3b9-9e87cf3c9ed0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.023649 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.023703 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/303e2849-4633-4722-b3b9-9e87cf3c9ed0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.023721 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm8j7\" (UniqueName: \"kubernetes.io/projected/303e2849-4633-4722-b3b9-9e87cf3c9ed0-kube-api-access-dm8j7\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.276427 4690 generic.go:334] "Generic (PLEG): container finished" podID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerID="be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d" exitCode=0 Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.276482 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtp8m" event={"ID":"303e2849-4633-4722-b3b9-9e87cf3c9ed0","Type":"ContainerDied","Data":"be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d"} Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.276545 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtp8m" event={"ID":"303e2849-4633-4722-b3b9-9e87cf3c9ed0","Type":"ContainerDied","Data":"e605bb586165d44fbf9700adbcb17f3266f2326c49d0eb747b4bf5430070878a"} Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.276551 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtp8m" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.276584 4690 scope.go:117] "RemoveContainer" containerID="be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.313265 4690 scope.go:117] "RemoveContainer" containerID="ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.337320 4690 scope.go:117] "RemoveContainer" containerID="847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.341921 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mtp8m"] Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.342107 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mtp8m"] Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.393988 4690 scope.go:117] "RemoveContainer" containerID="be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d" Dec 03 13:46:16 crc kubenswrapper[4690]: E1203 13:46:16.394415 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d\": container with ID starting with be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d not found: ID does not exist" containerID="be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.394451 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d"} err="failed to get container status \"be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d\": rpc error: code = NotFound desc = could not find container \"be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d\": container with ID starting with be14b685e09baa19687af138bd9d5806ddd7f80c93a03ca92d9508b9ef991f4d not found: ID does not exist" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.394475 4690 scope.go:117] "RemoveContainer" containerID="ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1" Dec 03 13:46:16 crc kubenswrapper[4690]: E1203 13:46:16.395200 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1\": container with ID starting with ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1 not found: ID does not exist" containerID="ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.395230 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1"} err="failed to get container status \"ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1\": rpc error: code = NotFound desc = could not find container \"ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1\": container with ID starting with ffc562df97b16ffdfd800222149d792ce66ce10dfdb78c2ab27be8e3f5b99fa1 not found: ID does not exist" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.395246 4690 scope.go:117] "RemoveContainer" containerID="847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd" Dec 03 13:46:16 crc kubenswrapper[4690]: E1203 13:46:16.395617 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd\": container with ID starting with 847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd not found: ID does not exist" containerID="847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd" Dec 03 13:46:16 crc kubenswrapper[4690]: I1203 13:46:16.395705 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd"} err="failed to get container status \"847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd\": rpc error: code = NotFound desc = could not find container \"847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd\": container with ID starting with 847e9833e8ebbcdf677b0bca5be2b8b35cb3d13dc08484555b09129dbe532acd not found: ID does not exist" Dec 03 13:46:18 crc kubenswrapper[4690]: I1203 13:46:18.329697 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" path="/var/lib/kubelet/pods/303e2849-4633-4722-b3b9-9e87cf3c9ed0/volumes" Dec 03 13:46:22 crc kubenswrapper[4690]: I1203 13:46:22.053182 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-q95w4"] Dec 03 13:46:22 crc kubenswrapper[4690]: I1203 13:46:22.086435 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-q95w4"] Dec 03 13:46:22 crc kubenswrapper[4690]: I1203 13:46:22.331978 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06684437-623f-4562-ac9e-55d65a10b540" path="/var/lib/kubelet/pods/06684437-623f-4562-ac9e-55d65a10b540/volumes" Dec 03 13:46:41 crc kubenswrapper[4690]: I1203 13:46:41.912038 4690 scope.go:117] "RemoveContainer" containerID="ef19ceea66a66fa06d0b1457c5a9722986dd5cc078ea0e9177614433aeaf2633" Dec 03 13:47:02 crc kubenswrapper[4690]: I1203 13:47:02.785953 4690 generic.go:334] "Generic (PLEG): container finished" podID="2f30c127-ad2e-4154-a344-61a21b57b3d0" containerID="6e0909f73af76e8ab0baf89a6043aabb38b5a6b628a96817732ff9399c435d3d" exitCode=0 Dec 03 13:47:02 crc kubenswrapper[4690]: I1203 13:47:02.786047 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" event={"ID":"2f30c127-ad2e-4154-a344-61a21b57b3d0","Type":"ContainerDied","Data":"6e0909f73af76e8ab0baf89a6043aabb38b5a6b628a96817732ff9399c435d3d"} Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.280130 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.389106 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhzcs\" (UniqueName: \"kubernetes.io/projected/2f30c127-ad2e-4154-a344-61a21b57b3d0-kube-api-access-hhzcs\") pod \"2f30c127-ad2e-4154-a344-61a21b57b3d0\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.389154 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-inventory\") pod \"2f30c127-ad2e-4154-a344-61a21b57b3d0\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.389355 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-ssh-key\") pod \"2f30c127-ad2e-4154-a344-61a21b57b3d0\" (UID: \"2f30c127-ad2e-4154-a344-61a21b57b3d0\") " Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.397173 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f30c127-ad2e-4154-a344-61a21b57b3d0-kube-api-access-hhzcs" (OuterVolumeSpecName: "kube-api-access-hhzcs") pod "2f30c127-ad2e-4154-a344-61a21b57b3d0" (UID: "2f30c127-ad2e-4154-a344-61a21b57b3d0"). InnerVolumeSpecName "kube-api-access-hhzcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.420945 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2f30c127-ad2e-4154-a344-61a21b57b3d0" (UID: "2f30c127-ad2e-4154-a344-61a21b57b3d0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.422944 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-inventory" (OuterVolumeSpecName: "inventory") pod "2f30c127-ad2e-4154-a344-61a21b57b3d0" (UID: "2f30c127-ad2e-4154-a344-61a21b57b3d0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.493088 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.493167 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhzcs\" (UniqueName: \"kubernetes.io/projected/2f30c127-ad2e-4154-a344-61a21b57b3d0-kube-api-access-hhzcs\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.493201 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f30c127-ad2e-4154-a344-61a21b57b3d0-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.829713 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" event={"ID":"2f30c127-ad2e-4154-a344-61a21b57b3d0","Type":"ContainerDied","Data":"4bc448cd2bb9c771c0bd2649963ce5a490f4ae72a4af2d72e66c3ef8706df027"} Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.830101 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bc448cd2bb9c771c0bd2649963ce5a490f4ae72a4af2d72e66c3ef8706df027" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.829908 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pksbh" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.909327 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh"] Dec 03 13:47:04 crc kubenswrapper[4690]: E1203 13:47:04.909856 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f30c127-ad2e-4154-a344-61a21b57b3d0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.909897 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f30c127-ad2e-4154-a344-61a21b57b3d0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 13:47:04 crc kubenswrapper[4690]: E1203 13:47:04.909915 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerName="registry-server" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.909921 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerName="registry-server" Dec 03 13:47:04 crc kubenswrapper[4690]: E1203 13:47:04.909938 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerName="extract-utilities" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.909944 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerName="extract-utilities" Dec 03 13:47:04 crc kubenswrapper[4690]: E1203 13:47:04.909967 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerName="extract-content" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.909974 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerName="extract-content" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.910190 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="303e2849-4633-4722-b3b9-9e87cf3c9ed0" containerName="registry-server" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.910204 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f30c127-ad2e-4154-a344-61a21b57b3d0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.911169 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.914484 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.914525 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.914680 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.920320 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh"] Dec 03 13:47:04 crc kubenswrapper[4690]: I1203 13:47:04.921468 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.004796 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssjxj\" (UniqueName: \"kubernetes.io/projected/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-kube-api-access-ssjxj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.005624 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.005956 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.108601 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssjxj\" (UniqueName: \"kubernetes.io/projected/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-kube-api-access-ssjxj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.109030 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.109269 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.115794 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.117559 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.131628 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssjxj\" (UniqueName: \"kubernetes.io/projected/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-kube-api-access-ssjxj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p55xh\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.231120 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.776225 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh"] Dec 03 13:47:05 crc kubenswrapper[4690]: I1203 13:47:05.839186 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" event={"ID":"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96","Type":"ContainerStarted","Data":"6402e2ada801b825d700952025b0f446b15822e0f1dc4280aaba7a577c40e83f"} Dec 03 13:47:06 crc kubenswrapper[4690]: I1203 13:47:06.849816 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" event={"ID":"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96","Type":"ContainerStarted","Data":"0f138e75d1e83b71e4a7df1b75c5b7a6b707fd9df06352ca6c079f000a066eb3"} Dec 03 13:47:06 crc kubenswrapper[4690]: I1203 13:47:06.871544 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" podStartSLOduration=2.354595269 podStartE2EDuration="2.871519839s" podCreationTimestamp="2025-12-03 13:47:04 +0000 UTC" firstStartedPulling="2025-12-03 13:47:05.781390265 +0000 UTC m=+2271.762310698" lastFinishedPulling="2025-12-03 13:47:06.298314845 +0000 UTC m=+2272.279235268" observedRunningTime="2025-12-03 13:47:06.869382815 +0000 UTC m=+2272.850303248" watchObservedRunningTime="2025-12-03 13:47:06.871519839 +0000 UTC m=+2272.852440282" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.112004 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ck48x"] Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.119387 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.126223 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ck48x"] Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.290510 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-catalog-content\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.290580 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhgpl\" (UniqueName: \"kubernetes.io/projected/927f188a-1022-40f0-b3da-11a67134134d-kube-api-access-jhgpl\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.290641 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-utilities\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.392849 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-catalog-content\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.392916 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhgpl\" (UniqueName: \"kubernetes.io/projected/927f188a-1022-40f0-b3da-11a67134134d-kube-api-access-jhgpl\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.392952 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-utilities\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.393411 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-utilities\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.393432 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-catalog-content\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.426052 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhgpl\" (UniqueName: \"kubernetes.io/projected/927f188a-1022-40f0-b3da-11a67134134d-kube-api-access-jhgpl\") pod \"redhat-marketplace-ck48x\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.444075 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:42 crc kubenswrapper[4690]: I1203 13:47:42.982036 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ck48x"] Dec 03 13:47:43 crc kubenswrapper[4690]: I1203 13:47:43.236569 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ck48x" event={"ID":"927f188a-1022-40f0-b3da-11a67134134d","Type":"ContainerStarted","Data":"bf26064a6fb771e29f8b9bd1a6a26c2266ac5d207d1c761ec2309af8ea50cec7"} Dec 03 13:47:43 crc kubenswrapper[4690]: I1203 13:47:43.906707 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5dnfq"] Dec 03 13:47:43 crc kubenswrapper[4690]: I1203 13:47:43.909911 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:43 crc kubenswrapper[4690]: I1203 13:47:43.920159 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5dnfq"] Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.035133 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzgkb\" (UniqueName: \"kubernetes.io/projected/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-kube-api-access-tzgkb\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.035219 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-catalog-content\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.035293 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-utilities\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.138187 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-catalog-content\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.138604 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-utilities\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.138637 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-catalog-content\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.138809 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzgkb\" (UniqueName: \"kubernetes.io/projected/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-kube-api-access-tzgkb\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.139131 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-utilities\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.159175 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzgkb\" (UniqueName: \"kubernetes.io/projected/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-kube-api-access-tzgkb\") pod \"redhat-operators-5dnfq\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.229491 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.258091 4690 generic.go:334] "Generic (PLEG): container finished" podID="927f188a-1022-40f0-b3da-11a67134134d" containerID="2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab" exitCode=0 Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.258156 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ck48x" event={"ID":"927f188a-1022-40f0-b3da-11a67134134d","Type":"ContainerDied","Data":"2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab"} Dec 03 13:47:44 crc kubenswrapper[4690]: I1203 13:47:44.762380 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5dnfq"] Dec 03 13:47:44 crc kubenswrapper[4690]: W1203 13:47:44.767132 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod597e2d20_d2f2_4b19_8e49_dbd3b49e5d0f.slice/crio-1bbe10704359125a7910000d702eb60adc381f0216c66aced44e369c37831dca WatchSource:0}: Error finding container 1bbe10704359125a7910000d702eb60adc381f0216c66aced44e369c37831dca: Status 404 returned error can't find the container with id 1bbe10704359125a7910000d702eb60adc381f0216c66aced44e369c37831dca Dec 03 13:47:45 crc kubenswrapper[4690]: I1203 13:47:45.269564 4690 generic.go:334] "Generic (PLEG): container finished" podID="927f188a-1022-40f0-b3da-11a67134134d" containerID="2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2" exitCode=0 Dec 03 13:47:45 crc kubenswrapper[4690]: I1203 13:47:45.271209 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ck48x" event={"ID":"927f188a-1022-40f0-b3da-11a67134134d","Type":"ContainerDied","Data":"2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2"} Dec 03 13:47:45 crc kubenswrapper[4690]: I1203 13:47:45.273663 4690 generic.go:334] "Generic (PLEG): container finished" podID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerID="be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948" exitCode=0 Dec 03 13:47:45 crc kubenswrapper[4690]: I1203 13:47:45.273713 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dnfq" event={"ID":"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f","Type":"ContainerDied","Data":"be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948"} Dec 03 13:47:45 crc kubenswrapper[4690]: I1203 13:47:45.273749 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dnfq" event={"ID":"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f","Type":"ContainerStarted","Data":"1bbe10704359125a7910000d702eb60adc381f0216c66aced44e369c37831dca"} Dec 03 13:47:46 crc kubenswrapper[4690]: I1203 13:47:46.287178 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ck48x" event={"ID":"927f188a-1022-40f0-b3da-11a67134134d","Type":"ContainerStarted","Data":"7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04"} Dec 03 13:47:46 crc kubenswrapper[4690]: I1203 13:47:46.317780 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ck48x" podStartSLOduration=2.883787238 podStartE2EDuration="4.317757462s" podCreationTimestamp="2025-12-03 13:47:42 +0000 UTC" firstStartedPulling="2025-12-03 13:47:44.261340096 +0000 UTC m=+2310.242260529" lastFinishedPulling="2025-12-03 13:47:45.69531032 +0000 UTC m=+2311.676230753" observedRunningTime="2025-12-03 13:47:46.309068444 +0000 UTC m=+2312.289988877" watchObservedRunningTime="2025-12-03 13:47:46.317757462 +0000 UTC m=+2312.298677895" Dec 03 13:47:47 crc kubenswrapper[4690]: I1203 13:47:47.318113 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dnfq" event={"ID":"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f","Type":"ContainerStarted","Data":"7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40"} Dec 03 13:47:49 crc kubenswrapper[4690]: I1203 13:47:49.339238 4690 generic.go:334] "Generic (PLEG): container finished" podID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerID="7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40" exitCode=0 Dec 03 13:47:49 crc kubenswrapper[4690]: I1203 13:47:49.339346 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dnfq" event={"ID":"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f","Type":"ContainerDied","Data":"7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40"} Dec 03 13:47:51 crc kubenswrapper[4690]: I1203 13:47:51.357954 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dnfq" event={"ID":"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f","Type":"ContainerStarted","Data":"6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d"} Dec 03 13:47:51 crc kubenswrapper[4690]: I1203 13:47:51.382163 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5dnfq" podStartSLOduration=2.778614594 podStartE2EDuration="8.382138525s" podCreationTimestamp="2025-12-03 13:47:43 +0000 UTC" firstStartedPulling="2025-12-03 13:47:45.274909126 +0000 UTC m=+2311.255829559" lastFinishedPulling="2025-12-03 13:47:50.878433057 +0000 UTC m=+2316.859353490" observedRunningTime="2025-12-03 13:47:51.374085602 +0000 UTC m=+2317.355006035" watchObservedRunningTime="2025-12-03 13:47:51.382138525 +0000 UTC m=+2317.363058968" Dec 03 13:47:52 crc kubenswrapper[4690]: I1203 13:47:52.444551 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:52 crc kubenswrapper[4690]: I1203 13:47:52.444930 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:52 crc kubenswrapper[4690]: I1203 13:47:52.492455 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:53 crc kubenswrapper[4690]: I1203 13:47:53.426571 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:53 crc kubenswrapper[4690]: I1203 13:47:53.698300 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ck48x"] Dec 03 13:47:54 crc kubenswrapper[4690]: I1203 13:47:54.230244 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:54 crc kubenswrapper[4690]: I1203 13:47:54.230304 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:47:55 crc kubenswrapper[4690]: I1203 13:47:55.289058 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5dnfq" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="registry-server" probeResult="failure" output=< Dec 03 13:47:55 crc kubenswrapper[4690]: timeout: failed to connect service ":50051" within 1s Dec 03 13:47:55 crc kubenswrapper[4690]: > Dec 03 13:47:55 crc kubenswrapper[4690]: I1203 13:47:55.396049 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ck48x" podUID="927f188a-1022-40f0-b3da-11a67134134d" containerName="registry-server" containerID="cri-o://7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04" gracePeriod=2 Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.386845 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.415242 4690 generic.go:334] "Generic (PLEG): container finished" podID="927f188a-1022-40f0-b3da-11a67134134d" containerID="7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04" exitCode=0 Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.415330 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ck48x" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.415345 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ck48x" event={"ID":"927f188a-1022-40f0-b3da-11a67134134d","Type":"ContainerDied","Data":"7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04"} Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.415774 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ck48x" event={"ID":"927f188a-1022-40f0-b3da-11a67134134d","Type":"ContainerDied","Data":"bf26064a6fb771e29f8b9bd1a6a26c2266ac5d207d1c761ec2309af8ea50cec7"} Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.415799 4690 scope.go:117] "RemoveContainer" containerID="7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.449513 4690 scope.go:117] "RemoveContainer" containerID="2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.470814 4690 scope.go:117] "RemoveContainer" containerID="2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.518137 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhgpl\" (UniqueName: \"kubernetes.io/projected/927f188a-1022-40f0-b3da-11a67134134d-kube-api-access-jhgpl\") pod \"927f188a-1022-40f0-b3da-11a67134134d\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.518366 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-catalog-content\") pod \"927f188a-1022-40f0-b3da-11a67134134d\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.518413 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-utilities\") pod \"927f188a-1022-40f0-b3da-11a67134134d\" (UID: \"927f188a-1022-40f0-b3da-11a67134134d\") " Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.520167 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-utilities" (OuterVolumeSpecName: "utilities") pod "927f188a-1022-40f0-b3da-11a67134134d" (UID: "927f188a-1022-40f0-b3da-11a67134134d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.531182 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/927f188a-1022-40f0-b3da-11a67134134d-kube-api-access-jhgpl" (OuterVolumeSpecName: "kube-api-access-jhgpl") pod "927f188a-1022-40f0-b3da-11a67134134d" (UID: "927f188a-1022-40f0-b3da-11a67134134d"). InnerVolumeSpecName "kube-api-access-jhgpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.532178 4690 scope.go:117] "RemoveContainer" containerID="7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04" Dec 03 13:47:56 crc kubenswrapper[4690]: E1203 13:47:56.532714 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04\": container with ID starting with 7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04 not found: ID does not exist" containerID="7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.532748 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04"} err="failed to get container status \"7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04\": rpc error: code = NotFound desc = could not find container \"7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04\": container with ID starting with 7569a99e8b57ab7983ace59a792b1a7e3db0d49f7f4e498b6801e7bdc3550a04 not found: ID does not exist" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.532774 4690 scope.go:117] "RemoveContainer" containerID="2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2" Dec 03 13:47:56 crc kubenswrapper[4690]: E1203 13:47:56.533103 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2\": container with ID starting with 2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2 not found: ID does not exist" containerID="2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.533130 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2"} err="failed to get container status \"2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2\": rpc error: code = NotFound desc = could not find container \"2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2\": container with ID starting with 2c16bddc611a10b90a1ce7453df0a6c62ea9d87261c2f99c182864d5f07398e2 not found: ID does not exist" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.533144 4690 scope.go:117] "RemoveContainer" containerID="2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab" Dec 03 13:47:56 crc kubenswrapper[4690]: E1203 13:47:56.533571 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab\": container with ID starting with 2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab not found: ID does not exist" containerID="2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.533595 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab"} err="failed to get container status \"2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab\": rpc error: code = NotFound desc = could not find container \"2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab\": container with ID starting with 2fed4c7ec1d89f88029af3432c90b5f23fd4c659eba2f3f60389ce428eddb9ab not found: ID does not exist" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.539292 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "927f188a-1022-40f0-b3da-11a67134134d" (UID: "927f188a-1022-40f0-b3da-11a67134134d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.621597 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.621640 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927f188a-1022-40f0-b3da-11a67134134d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.621653 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhgpl\" (UniqueName: \"kubernetes.io/projected/927f188a-1022-40f0-b3da-11a67134134d-kube-api-access-jhgpl\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.750733 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ck48x"] Dec 03 13:47:56 crc kubenswrapper[4690]: I1203 13:47:56.758926 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ck48x"] Dec 03 13:47:58 crc kubenswrapper[4690]: I1203 13:47:58.329987 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="927f188a-1022-40f0-b3da-11a67134134d" path="/var/lib/kubelet/pods/927f188a-1022-40f0-b3da-11a67134134d/volumes" Dec 03 13:48:04 crc kubenswrapper[4690]: I1203 13:48:04.276162 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:48:04 crc kubenswrapper[4690]: I1203 13:48:04.334087 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:48:04 crc kubenswrapper[4690]: I1203 13:48:04.517230 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5dnfq"] Dec 03 13:48:05 crc kubenswrapper[4690]: I1203 13:48:05.546696 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5dnfq" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="registry-server" containerID="cri-o://6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d" gracePeriod=2 Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.046272 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.164319 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-utilities\") pod \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.164463 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzgkb\" (UniqueName: \"kubernetes.io/projected/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-kube-api-access-tzgkb\") pod \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.164531 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-catalog-content\") pod \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\" (UID: \"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f\") " Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.166224 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-utilities" (OuterVolumeSpecName: "utilities") pod "597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" (UID: "597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.171008 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-kube-api-access-tzgkb" (OuterVolumeSpecName: "kube-api-access-tzgkb") pod "597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" (UID: "597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f"). InnerVolumeSpecName "kube-api-access-tzgkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.267272 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.267313 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzgkb\" (UniqueName: \"kubernetes.io/projected/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-kube-api-access-tzgkb\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.282686 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" (UID: "597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.369228 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.557153 4690 generic.go:334] "Generic (PLEG): container finished" podID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerID="6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d" exitCode=0 Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.557207 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dnfq" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.557226 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dnfq" event={"ID":"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f","Type":"ContainerDied","Data":"6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d"} Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.558816 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dnfq" event={"ID":"597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f","Type":"ContainerDied","Data":"1bbe10704359125a7910000d702eb60adc381f0216c66aced44e369c37831dca"} Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.558839 4690 scope.go:117] "RemoveContainer" containerID="6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.593522 4690 scope.go:117] "RemoveContainer" containerID="7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.595574 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5dnfq"] Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.605395 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5dnfq"] Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.621660 4690 scope.go:117] "RemoveContainer" containerID="be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.667035 4690 scope.go:117] "RemoveContainer" containerID="6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d" Dec 03 13:48:06 crc kubenswrapper[4690]: E1203 13:48:06.668082 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d\": container with ID starting with 6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d not found: ID does not exist" containerID="6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.668126 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d"} err="failed to get container status \"6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d\": rpc error: code = NotFound desc = could not find container \"6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d\": container with ID starting with 6055e5e703528427c1d282c8888e9e3f27b7427c85af3d5a75948c6fe6ee884d not found: ID does not exist" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.668152 4690 scope.go:117] "RemoveContainer" containerID="7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40" Dec 03 13:48:06 crc kubenswrapper[4690]: E1203 13:48:06.668382 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40\": container with ID starting with 7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40 not found: ID does not exist" containerID="7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.668407 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40"} err="failed to get container status \"7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40\": rpc error: code = NotFound desc = could not find container \"7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40\": container with ID starting with 7b36c29dfb0c0ebeffd644e9538c17e2c3c60c91209bb5f14487feb2e7e31a40 not found: ID does not exist" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.668424 4690 scope.go:117] "RemoveContainer" containerID="be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948" Dec 03 13:48:06 crc kubenswrapper[4690]: E1203 13:48:06.668661 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948\": container with ID starting with be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948 not found: ID does not exist" containerID="be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948" Dec 03 13:48:06 crc kubenswrapper[4690]: I1203 13:48:06.668698 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948"} err="failed to get container status \"be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948\": rpc error: code = NotFound desc = could not find container \"be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948\": container with ID starting with be8beeb3581cc43411fa25d6b4d9fd3d6b65fce5ac5b337af1e19a7dbdb81948 not found: ID does not exist" Dec 03 13:48:08 crc kubenswrapper[4690]: I1203 13:48:08.347247 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" path="/var/lib/kubelet/pods/597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f/volumes" Dec 03 13:48:16 crc kubenswrapper[4690]: I1203 13:48:16.824006 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:48:16 crc kubenswrapper[4690]: I1203 13:48:16.824667 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:48:46 crc kubenswrapper[4690]: I1203 13:48:46.823796 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:48:46 crc kubenswrapper[4690]: I1203 13:48:46.824584 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:48:54 crc kubenswrapper[4690]: I1203 13:48:54.018455 4690 generic.go:334] "Generic (PLEG): container finished" podID="69f0b3dc-dc18-4a71-ad0c-a5023db9ca96" containerID="0f138e75d1e83b71e4a7df1b75c5b7a6b707fd9df06352ca6c079f000a066eb3" exitCode=0 Dec 03 13:48:54 crc kubenswrapper[4690]: I1203 13:48:54.018565 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" event={"ID":"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96","Type":"ContainerDied","Data":"0f138e75d1e83b71e4a7df1b75c5b7a6b707fd9df06352ca6c079f000a066eb3"} Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.487887 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.622414 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-ssh-key\") pod \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.623999 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssjxj\" (UniqueName: \"kubernetes.io/projected/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-kube-api-access-ssjxj\") pod \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.628995 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-inventory\") pod \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\" (UID: \"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96\") " Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.630267 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-kube-api-access-ssjxj" (OuterVolumeSpecName: "kube-api-access-ssjxj") pod "69f0b3dc-dc18-4a71-ad0c-a5023db9ca96" (UID: "69f0b3dc-dc18-4a71-ad0c-a5023db9ca96"). InnerVolumeSpecName "kube-api-access-ssjxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.655237 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-inventory" (OuterVolumeSpecName: "inventory") pod "69f0b3dc-dc18-4a71-ad0c-a5023db9ca96" (UID: "69f0b3dc-dc18-4a71-ad0c-a5023db9ca96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.665724 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "69f0b3dc-dc18-4a71-ad0c-a5023db9ca96" (UID: "69f0b3dc-dc18-4a71-ad0c-a5023db9ca96"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.731672 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.731898 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssjxj\" (UniqueName: \"kubernetes.io/projected/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-kube-api-access-ssjxj\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:55 crc kubenswrapper[4690]: I1203 13:48:55.731982 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69f0b3dc-dc18-4a71-ad0c-a5023db9ca96-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.064811 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" event={"ID":"69f0b3dc-dc18-4a71-ad0c-a5023db9ca96","Type":"ContainerDied","Data":"6402e2ada801b825d700952025b0f446b15822e0f1dc4280aaba7a577c40e83f"} Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.064906 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6402e2ada801b825d700952025b0f446b15822e0f1dc4280aaba7a577c40e83f" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.065011 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p55xh" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.134401 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr"] Dec 03 13:48:56 crc kubenswrapper[4690]: E1203 13:48:56.134964 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f0b3dc-dc18-4a71-ad0c-a5023db9ca96" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.134987 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f0b3dc-dc18-4a71-ad0c-a5023db9ca96" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:48:56 crc kubenswrapper[4690]: E1203 13:48:56.135016 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927f188a-1022-40f0-b3da-11a67134134d" containerName="extract-utilities" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135025 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="927f188a-1022-40f0-b3da-11a67134134d" containerName="extract-utilities" Dec 03 13:48:56 crc kubenswrapper[4690]: E1203 13:48:56.135037 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927f188a-1022-40f0-b3da-11a67134134d" containerName="extract-content" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135046 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="927f188a-1022-40f0-b3da-11a67134134d" containerName="extract-content" Dec 03 13:48:56 crc kubenswrapper[4690]: E1203 13:48:56.135061 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="extract-utilities" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135071 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="extract-utilities" Dec 03 13:48:56 crc kubenswrapper[4690]: E1203 13:48:56.135082 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927f188a-1022-40f0-b3da-11a67134134d" containerName="registry-server" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135089 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="927f188a-1022-40f0-b3da-11a67134134d" containerName="registry-server" Dec 03 13:48:56 crc kubenswrapper[4690]: E1203 13:48:56.135117 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="registry-server" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135126 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="registry-server" Dec 03 13:48:56 crc kubenswrapper[4690]: E1203 13:48:56.135138 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="extract-content" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135145 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="extract-content" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135388 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f0b3dc-dc18-4a71-ad0c-a5023db9ca96" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135415 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="597e2d20-d2f2-4b19-8e49-dbd3b49e5d0f" containerName="registry-server" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.135425 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="927f188a-1022-40f0-b3da-11a67134134d" containerName="registry-server" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.136699 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.139206 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.139498 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.140124 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.140479 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.144659 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr"] Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.247251 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d292l\" (UniqueName: \"kubernetes.io/projected/96b45681-d5bf-474b-ac2c-2f8df92494d9-kube-api-access-d292l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.247379 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.247527 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.350166 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.350266 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d292l\" (UniqueName: \"kubernetes.io/projected/96b45681-d5bf-474b-ac2c-2f8df92494d9-kube-api-access-d292l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.350331 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.354270 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.360456 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.374765 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d292l\" (UniqueName: \"kubernetes.io/projected/96b45681-d5bf-474b-ac2c-2f8df92494d9-kube-api-access-d292l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:56 crc kubenswrapper[4690]: I1203 13:48:56.455686 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:48:57 crc kubenswrapper[4690]: I1203 13:48:57.001110 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr"] Dec 03 13:48:57 crc kubenswrapper[4690]: W1203 13:48:57.003205 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96b45681_d5bf_474b_ac2c_2f8df92494d9.slice/crio-94338a18ea08f210abbdd93eeca9d4728d650cc84ff8c0a14cb89c964f4c3629 WatchSource:0}: Error finding container 94338a18ea08f210abbdd93eeca9d4728d650cc84ff8c0a14cb89c964f4c3629: Status 404 returned error can't find the container with id 94338a18ea08f210abbdd93eeca9d4728d650cc84ff8c0a14cb89c964f4c3629 Dec 03 13:48:57 crc kubenswrapper[4690]: I1203 13:48:57.077540 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" event={"ID":"96b45681-d5bf-474b-ac2c-2f8df92494d9","Type":"ContainerStarted","Data":"94338a18ea08f210abbdd93eeca9d4728d650cc84ff8c0a14cb89c964f4c3629"} Dec 03 13:48:59 crc kubenswrapper[4690]: I1203 13:48:59.099096 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" event={"ID":"96b45681-d5bf-474b-ac2c-2f8df92494d9","Type":"ContainerStarted","Data":"e96804ad70a380d059ed2103c6654beea4b231cdc52c7d080fd90f753a735328"} Dec 03 13:48:59 crc kubenswrapper[4690]: I1203 13:48:59.116110 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" podStartSLOduration=2.047363266 podStartE2EDuration="3.116091042s" podCreationTimestamp="2025-12-03 13:48:56 +0000 UTC" firstStartedPulling="2025-12-03 13:48:57.017471556 +0000 UTC m=+2382.998391989" lastFinishedPulling="2025-12-03 13:48:58.086199332 +0000 UTC m=+2384.067119765" observedRunningTime="2025-12-03 13:48:59.113717063 +0000 UTC m=+2385.094637506" watchObservedRunningTime="2025-12-03 13:48:59.116091042 +0000 UTC m=+2385.097011475" Dec 03 13:49:04 crc kubenswrapper[4690]: I1203 13:49:04.146047 4690 generic.go:334] "Generic (PLEG): container finished" podID="96b45681-d5bf-474b-ac2c-2f8df92494d9" containerID="e96804ad70a380d059ed2103c6654beea4b231cdc52c7d080fd90f753a735328" exitCode=0 Dec 03 13:49:04 crc kubenswrapper[4690]: I1203 13:49:04.146131 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" event={"ID":"96b45681-d5bf-474b-ac2c-2f8df92494d9","Type":"ContainerDied","Data":"e96804ad70a380d059ed2103c6654beea4b231cdc52c7d080fd90f753a735328"} Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.604472 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.673854 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-ssh-key\") pod \"96b45681-d5bf-474b-ac2c-2f8df92494d9\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.674368 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-inventory\") pod \"96b45681-d5bf-474b-ac2c-2f8df92494d9\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.674587 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d292l\" (UniqueName: \"kubernetes.io/projected/96b45681-d5bf-474b-ac2c-2f8df92494d9-kube-api-access-d292l\") pod \"96b45681-d5bf-474b-ac2c-2f8df92494d9\" (UID: \"96b45681-d5bf-474b-ac2c-2f8df92494d9\") " Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.681299 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b45681-d5bf-474b-ac2c-2f8df92494d9-kube-api-access-d292l" (OuterVolumeSpecName: "kube-api-access-d292l") pod "96b45681-d5bf-474b-ac2c-2f8df92494d9" (UID: "96b45681-d5bf-474b-ac2c-2f8df92494d9"). InnerVolumeSpecName "kube-api-access-d292l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.704126 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-inventory" (OuterVolumeSpecName: "inventory") pod "96b45681-d5bf-474b-ac2c-2f8df92494d9" (UID: "96b45681-d5bf-474b-ac2c-2f8df92494d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.704511 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "96b45681-d5bf-474b-ac2c-2f8df92494d9" (UID: "96b45681-d5bf-474b-ac2c-2f8df92494d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.777374 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d292l\" (UniqueName: \"kubernetes.io/projected/96b45681-d5bf-474b-ac2c-2f8df92494d9-kube-api-access-d292l\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.777769 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:05 crc kubenswrapper[4690]: I1203 13:49:05.777781 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96b45681-d5bf-474b-ac2c-2f8df92494d9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.181249 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" event={"ID":"96b45681-d5bf-474b-ac2c-2f8df92494d9","Type":"ContainerDied","Data":"94338a18ea08f210abbdd93eeca9d4728d650cc84ff8c0a14cb89c964f4c3629"} Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.181530 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94338a18ea08f210abbdd93eeca9d4728d650cc84ff8c0a14cb89c964f4c3629" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.181336 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.246960 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb"] Dec 03 13:49:06 crc kubenswrapper[4690]: E1203 13:49:06.247783 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96b45681-d5bf-474b-ac2c-2f8df92494d9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.247811 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="96b45681-d5bf-474b-ac2c-2f8df92494d9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.248078 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="96b45681-d5bf-474b-ac2c-2f8df92494d9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.249064 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.252531 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.252540 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.253649 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.253840 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.261617 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb"] Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.391136 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8pc5\" (UniqueName: \"kubernetes.io/projected/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-kube-api-access-w8pc5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.391204 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.391390 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.492895 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.493032 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8pc5\" (UniqueName: \"kubernetes.io/projected/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-kube-api-access-w8pc5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.493081 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.499029 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.501416 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.510841 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8pc5\" (UniqueName: \"kubernetes.io/projected/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-kube-api-access-w8pc5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kglnb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:06 crc kubenswrapper[4690]: I1203 13:49:06.570433 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:07 crc kubenswrapper[4690]: W1203 13:49:07.084721 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde573b60_ca9e_483c_8b4a_6d62fc6f56eb.slice/crio-72bfa3a93adbd2055382cf9d0090e3e85cbd7aee19945a7bbdd0ec73714034a9 WatchSource:0}: Error finding container 72bfa3a93adbd2055382cf9d0090e3e85cbd7aee19945a7bbdd0ec73714034a9: Status 404 returned error can't find the container with id 72bfa3a93adbd2055382cf9d0090e3e85cbd7aee19945a7bbdd0ec73714034a9 Dec 03 13:49:07 crc kubenswrapper[4690]: I1203 13:49:07.085494 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb"] Dec 03 13:49:07 crc kubenswrapper[4690]: I1203 13:49:07.089768 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:49:07 crc kubenswrapper[4690]: I1203 13:49:07.192404 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" event={"ID":"de573b60-ca9e-483c-8b4a-6d62fc6f56eb","Type":"ContainerStarted","Data":"72bfa3a93adbd2055382cf9d0090e3e85cbd7aee19945a7bbdd0ec73714034a9"} Dec 03 13:49:08 crc kubenswrapper[4690]: I1203 13:49:08.205466 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" event={"ID":"de573b60-ca9e-483c-8b4a-6d62fc6f56eb","Type":"ContainerStarted","Data":"9f6fc12d92c08ef6215e6d6f57cf963d2295a536429e43a807773cc5e1654c7d"} Dec 03 13:49:08 crc kubenswrapper[4690]: I1203 13:49:08.232183 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" podStartSLOduration=1.611813432 podStartE2EDuration="2.232158581s" podCreationTimestamp="2025-12-03 13:49:06 +0000 UTC" firstStartedPulling="2025-12-03 13:49:07.089557838 +0000 UTC m=+2393.070478271" lastFinishedPulling="2025-12-03 13:49:07.709902987 +0000 UTC m=+2393.690823420" observedRunningTime="2025-12-03 13:49:08.223092243 +0000 UTC m=+2394.204012676" watchObservedRunningTime="2025-12-03 13:49:08.232158581 +0000 UTC m=+2394.213079014" Dec 03 13:49:16 crc kubenswrapper[4690]: I1203 13:49:16.823600 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:49:16 crc kubenswrapper[4690]: I1203 13:49:16.824247 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:49:16 crc kubenswrapper[4690]: I1203 13:49:16.824294 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:49:16 crc kubenswrapper[4690]: I1203 13:49:16.825145 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:49:16 crc kubenswrapper[4690]: I1203 13:49:16.825202 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" gracePeriod=600 Dec 03 13:49:16 crc kubenswrapper[4690]: E1203 13:49:16.949654 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:49:17 crc kubenswrapper[4690]: I1203 13:49:17.300390 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" exitCode=0 Dec 03 13:49:17 crc kubenswrapper[4690]: I1203 13:49:17.300440 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1"} Dec 03 13:49:17 crc kubenswrapper[4690]: I1203 13:49:17.300490 4690 scope.go:117] "RemoveContainer" containerID="4587129f3424bae3d234d695d33b32c949d4b82d91b7d56a0e8ebc65700b9dec" Dec 03 13:49:17 crc kubenswrapper[4690]: I1203 13:49:17.301401 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:49:17 crc kubenswrapper[4690]: E1203 13:49:17.301689 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:49:30 crc kubenswrapper[4690]: I1203 13:49:30.315019 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:49:30 crc kubenswrapper[4690]: E1203 13:49:30.316007 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:49:42 crc kubenswrapper[4690]: I1203 13:49:42.316113 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:49:42 crc kubenswrapper[4690]: E1203 13:49:42.317222 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:49:46 crc kubenswrapper[4690]: I1203 13:49:46.608628 4690 generic.go:334] "Generic (PLEG): container finished" podID="de573b60-ca9e-483c-8b4a-6d62fc6f56eb" containerID="9f6fc12d92c08ef6215e6d6f57cf963d2295a536429e43a807773cc5e1654c7d" exitCode=0 Dec 03 13:49:46 crc kubenswrapper[4690]: I1203 13:49:46.608717 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" event={"ID":"de573b60-ca9e-483c-8b4a-6d62fc6f56eb","Type":"ContainerDied","Data":"9f6fc12d92c08ef6215e6d6f57cf963d2295a536429e43a807773cc5e1654c7d"} Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.082742 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.169921 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8pc5\" (UniqueName: \"kubernetes.io/projected/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-kube-api-access-w8pc5\") pod \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.169995 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-inventory\") pod \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.170166 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-ssh-key\") pod \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\" (UID: \"de573b60-ca9e-483c-8b4a-6d62fc6f56eb\") " Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.178550 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-kube-api-access-w8pc5" (OuterVolumeSpecName: "kube-api-access-w8pc5") pod "de573b60-ca9e-483c-8b4a-6d62fc6f56eb" (UID: "de573b60-ca9e-483c-8b4a-6d62fc6f56eb"). InnerVolumeSpecName "kube-api-access-w8pc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.220061 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de573b60-ca9e-483c-8b4a-6d62fc6f56eb" (UID: "de573b60-ca9e-483c-8b4a-6d62fc6f56eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.220334 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-inventory" (OuterVolumeSpecName: "inventory") pod "de573b60-ca9e-483c-8b4a-6d62fc6f56eb" (UID: "de573b60-ca9e-483c-8b4a-6d62fc6f56eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.272848 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8pc5\" (UniqueName: \"kubernetes.io/projected/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-kube-api-access-w8pc5\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.272916 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.272930 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de573b60-ca9e-483c-8b4a-6d62fc6f56eb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.627531 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" event={"ID":"de573b60-ca9e-483c-8b4a-6d62fc6f56eb","Type":"ContainerDied","Data":"72bfa3a93adbd2055382cf9d0090e3e85cbd7aee19945a7bbdd0ec73714034a9"} Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.627912 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72bfa3a93adbd2055382cf9d0090e3e85cbd7aee19945a7bbdd0ec73714034a9" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.627627 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kglnb" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.823993 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl"] Dec 03 13:49:48 crc kubenswrapper[4690]: E1203 13:49:48.824529 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de573b60-ca9e-483c-8b4a-6d62fc6f56eb" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.824548 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="de573b60-ca9e-483c-8b4a-6d62fc6f56eb" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.824786 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="de573b60-ca9e-483c-8b4a-6d62fc6f56eb" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.825611 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.830081 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.830167 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.830218 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.830332 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.834213 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl"] Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.991102 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.991209 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:48 crc kubenswrapper[4690]: I1203 13:49:48.991259 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w8rp\" (UniqueName: \"kubernetes.io/projected/ede30151-25e0-47e4-98b4-bf7ca4d55255-kube-api-access-4w8rp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:49 crc kubenswrapper[4690]: I1203 13:49:49.092907 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:49 crc kubenswrapper[4690]: I1203 13:49:49.093000 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:49 crc kubenswrapper[4690]: I1203 13:49:49.093050 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w8rp\" (UniqueName: \"kubernetes.io/projected/ede30151-25e0-47e4-98b4-bf7ca4d55255-kube-api-access-4w8rp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:49 crc kubenswrapper[4690]: I1203 13:49:49.098651 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:49 crc kubenswrapper[4690]: I1203 13:49:49.101239 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:49 crc kubenswrapper[4690]: I1203 13:49:49.112887 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w8rp\" (UniqueName: \"kubernetes.io/projected/ede30151-25e0-47e4-98b4-bf7ca4d55255-kube-api-access-4w8rp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-77rvl\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:49 crc kubenswrapper[4690]: I1203 13:49:49.144759 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:49:49 crc kubenswrapper[4690]: I1203 13:49:49.734919 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl"] Dec 03 13:49:50 crc kubenswrapper[4690]: I1203 13:49:50.647365 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" event={"ID":"ede30151-25e0-47e4-98b4-bf7ca4d55255","Type":"ContainerStarted","Data":"ad1c7e779467a013e7882f13003ebd70f710f435adeb4b0a54c4ae7cfc543de2"} Dec 03 13:49:50 crc kubenswrapper[4690]: I1203 13:49:50.647738 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" event={"ID":"ede30151-25e0-47e4-98b4-bf7ca4d55255","Type":"ContainerStarted","Data":"e1228a63ccd8540a49703c9382e3114de00e8cc065843d51616f1da40917135f"} Dec 03 13:49:50 crc kubenswrapper[4690]: I1203 13:49:50.670475 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" podStartSLOduration=2.260161505 podStartE2EDuration="2.670456352s" podCreationTimestamp="2025-12-03 13:49:48 +0000 UTC" firstStartedPulling="2025-12-03 13:49:49.746404202 +0000 UTC m=+2435.727324635" lastFinishedPulling="2025-12-03 13:49:50.156699049 +0000 UTC m=+2436.137619482" observedRunningTime="2025-12-03 13:49:50.665586489 +0000 UTC m=+2436.646506922" watchObservedRunningTime="2025-12-03 13:49:50.670456352 +0000 UTC m=+2436.651376785" Dec 03 13:49:53 crc kubenswrapper[4690]: I1203 13:49:53.315731 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:49:53 crc kubenswrapper[4690]: E1203 13:49:53.316442 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:50:05 crc kubenswrapper[4690]: I1203 13:50:05.315068 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:50:05 crc kubenswrapper[4690]: E1203 13:50:05.315934 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:50:16 crc kubenswrapper[4690]: I1203 13:50:16.315305 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:50:16 crc kubenswrapper[4690]: E1203 13:50:16.316137 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:50:27 crc kubenswrapper[4690]: I1203 13:50:27.315305 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:50:27 crc kubenswrapper[4690]: E1203 13:50:27.316273 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:50:42 crc kubenswrapper[4690]: I1203 13:50:42.315654 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:50:42 crc kubenswrapper[4690]: E1203 13:50:42.316536 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:50:43 crc kubenswrapper[4690]: I1203 13:50:43.304885 4690 generic.go:334] "Generic (PLEG): container finished" podID="ede30151-25e0-47e4-98b4-bf7ca4d55255" containerID="ad1c7e779467a013e7882f13003ebd70f710f435adeb4b0a54c4ae7cfc543de2" exitCode=0 Dec 03 13:50:43 crc kubenswrapper[4690]: I1203 13:50:43.305315 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" event={"ID":"ede30151-25e0-47e4-98b4-bf7ca4d55255","Type":"ContainerDied","Data":"ad1c7e779467a013e7882f13003ebd70f710f435adeb4b0a54c4ae7cfc543de2"} Dec 03 13:50:44 crc kubenswrapper[4690]: I1203 13:50:44.768909 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:50:44 crc kubenswrapper[4690]: I1203 13:50:44.949045 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-inventory\") pod \"ede30151-25e0-47e4-98b4-bf7ca4d55255\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " Dec 03 13:50:44 crc kubenswrapper[4690]: I1203 13:50:44.949240 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w8rp\" (UniqueName: \"kubernetes.io/projected/ede30151-25e0-47e4-98b4-bf7ca4d55255-kube-api-access-4w8rp\") pod \"ede30151-25e0-47e4-98b4-bf7ca4d55255\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " Dec 03 13:50:44 crc kubenswrapper[4690]: I1203 13:50:44.949432 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-ssh-key\") pod \"ede30151-25e0-47e4-98b4-bf7ca4d55255\" (UID: \"ede30151-25e0-47e4-98b4-bf7ca4d55255\") " Dec 03 13:50:44 crc kubenswrapper[4690]: I1203 13:50:44.956586 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ede30151-25e0-47e4-98b4-bf7ca4d55255-kube-api-access-4w8rp" (OuterVolumeSpecName: "kube-api-access-4w8rp") pod "ede30151-25e0-47e4-98b4-bf7ca4d55255" (UID: "ede30151-25e0-47e4-98b4-bf7ca4d55255"). InnerVolumeSpecName "kube-api-access-4w8rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:50:44 crc kubenswrapper[4690]: I1203 13:50:44.985672 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-inventory" (OuterVolumeSpecName: "inventory") pod "ede30151-25e0-47e4-98b4-bf7ca4d55255" (UID: "ede30151-25e0-47e4-98b4-bf7ca4d55255"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:50:44 crc kubenswrapper[4690]: I1203 13:50:44.986856 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ede30151-25e0-47e4-98b4-bf7ca4d55255" (UID: "ede30151-25e0-47e4-98b4-bf7ca4d55255"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.052122 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w8rp\" (UniqueName: \"kubernetes.io/projected/ede30151-25e0-47e4-98b4-bf7ca4d55255-kube-api-access-4w8rp\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.052395 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.052566 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede30151-25e0-47e4-98b4-bf7ca4d55255-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.329847 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" event={"ID":"ede30151-25e0-47e4-98b4-bf7ca4d55255","Type":"ContainerDied","Data":"e1228a63ccd8540a49703c9382e3114de00e8cc065843d51616f1da40917135f"} Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.329927 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1228a63ccd8540a49703c9382e3114de00e8cc065843d51616f1da40917135f" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.329939 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-77rvl" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.419667 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gvv8v"] Dec 03 13:50:45 crc kubenswrapper[4690]: E1203 13:50:45.420470 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede30151-25e0-47e4-98b4-bf7ca4d55255" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.420579 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede30151-25e0-47e4-98b4-bf7ca4d55255" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.420912 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="ede30151-25e0-47e4-98b4-bf7ca4d55255" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.421730 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.426156 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.426599 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.426790 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.427393 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.430294 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gvv8v"] Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.564073 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grtw8\" (UniqueName: \"kubernetes.io/projected/91d255ec-88b0-416f-a990-92d930335b46-kube-api-access-grtw8\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.564575 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.564611 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.666294 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.666764 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.666943 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grtw8\" (UniqueName: \"kubernetes.io/projected/91d255ec-88b0-416f-a990-92d930335b46-kube-api-access-grtw8\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.674245 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.675485 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.691691 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grtw8\" (UniqueName: \"kubernetes.io/projected/91d255ec-88b0-416f-a990-92d930335b46-kube-api-access-grtw8\") pod \"ssh-known-hosts-edpm-deployment-gvv8v\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:45 crc kubenswrapper[4690]: I1203 13:50:45.751116 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:46 crc kubenswrapper[4690]: I1203 13:50:46.286592 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gvv8v"] Dec 03 13:50:46 crc kubenswrapper[4690]: I1203 13:50:46.338635 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" event={"ID":"91d255ec-88b0-416f-a990-92d930335b46","Type":"ContainerStarted","Data":"206de4a7957f0bec0d095f389e02011e82fc4ee0deeeaaa52b6fd8bb7b1f8e4c"} Dec 03 13:50:47 crc kubenswrapper[4690]: I1203 13:50:47.350482 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" event={"ID":"91d255ec-88b0-416f-a990-92d930335b46","Type":"ContainerStarted","Data":"63868e7cf25440f7ad8e41fc0ef737f7e292f9813fa04f434f7a85088552e584"} Dec 03 13:50:47 crc kubenswrapper[4690]: I1203 13:50:47.369581 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" podStartSLOduration=1.7446755980000002 podStartE2EDuration="2.369554737s" podCreationTimestamp="2025-12-03 13:50:45 +0000 UTC" firstStartedPulling="2025-12-03 13:50:46.28908308 +0000 UTC m=+2492.270003513" lastFinishedPulling="2025-12-03 13:50:46.913962209 +0000 UTC m=+2492.894882652" observedRunningTime="2025-12-03 13:50:47.368093971 +0000 UTC m=+2493.349014434" watchObservedRunningTime="2025-12-03 13:50:47.369554737 +0000 UTC m=+2493.350475170" Dec 03 13:50:54 crc kubenswrapper[4690]: I1203 13:50:54.324662 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:50:54 crc kubenswrapper[4690]: E1203 13:50:54.325742 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:50:54 crc kubenswrapper[4690]: I1203 13:50:54.426548 4690 generic.go:334] "Generic (PLEG): container finished" podID="91d255ec-88b0-416f-a990-92d930335b46" containerID="63868e7cf25440f7ad8e41fc0ef737f7e292f9813fa04f434f7a85088552e584" exitCode=0 Dec 03 13:50:54 crc kubenswrapper[4690]: I1203 13:50:54.426598 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" event={"ID":"91d255ec-88b0-416f-a990-92d930335b46","Type":"ContainerDied","Data":"63868e7cf25440f7ad8e41fc0ef737f7e292f9813fa04f434f7a85088552e584"} Dec 03 13:50:55 crc kubenswrapper[4690]: I1203 13:50:55.944088 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.095713 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-inventory-0\") pod \"91d255ec-88b0-416f-a990-92d930335b46\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.095774 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-ssh-key-openstack-edpm-ipam\") pod \"91d255ec-88b0-416f-a990-92d930335b46\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.095931 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grtw8\" (UniqueName: \"kubernetes.io/projected/91d255ec-88b0-416f-a990-92d930335b46-kube-api-access-grtw8\") pod \"91d255ec-88b0-416f-a990-92d930335b46\" (UID: \"91d255ec-88b0-416f-a990-92d930335b46\") " Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.114146 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91d255ec-88b0-416f-a990-92d930335b46-kube-api-access-grtw8" (OuterVolumeSpecName: "kube-api-access-grtw8") pod "91d255ec-88b0-416f-a990-92d930335b46" (UID: "91d255ec-88b0-416f-a990-92d930335b46"). InnerVolumeSpecName "kube-api-access-grtw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.129556 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "91d255ec-88b0-416f-a990-92d930335b46" (UID: "91d255ec-88b0-416f-a990-92d930335b46"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.129904 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "91d255ec-88b0-416f-a990-92d930335b46" (UID: "91d255ec-88b0-416f-a990-92d930335b46"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.198937 4690 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.198982 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91d255ec-88b0-416f-a990-92d930335b46-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.198994 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grtw8\" (UniqueName: \"kubernetes.io/projected/91d255ec-88b0-416f-a990-92d930335b46-kube-api-access-grtw8\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.452456 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" event={"ID":"91d255ec-88b0-416f-a990-92d930335b46","Type":"ContainerDied","Data":"206de4a7957f0bec0d095f389e02011e82fc4ee0deeeaaa52b6fd8bb7b1f8e4c"} Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.452499 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="206de4a7957f0bec0d095f389e02011e82fc4ee0deeeaaa52b6fd8bb7b1f8e4c" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.452958 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gvv8v" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.523913 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4"] Dec 03 13:50:56 crc kubenswrapper[4690]: E1203 13:50:56.524442 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d255ec-88b0-416f-a990-92d930335b46" containerName="ssh-known-hosts-edpm-deployment" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.524465 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d255ec-88b0-416f-a990-92d930335b46" containerName="ssh-known-hosts-edpm-deployment" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.524682 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d255ec-88b0-416f-a990-92d930335b46" containerName="ssh-known-hosts-edpm-deployment" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.525502 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.527471 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.527892 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.528257 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.528420 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.536698 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4"] Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.608211 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.608301 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jxv7\" (UniqueName: \"kubernetes.io/projected/80400476-b39e-4e4f-b57f-8813951d0f1a-kube-api-access-8jxv7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.608442 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.710500 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.710684 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.710729 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jxv7\" (UniqueName: \"kubernetes.io/projected/80400476-b39e-4e4f-b57f-8813951d0f1a-kube-api-access-8jxv7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.717238 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.718485 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.730627 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jxv7\" (UniqueName: \"kubernetes.io/projected/80400476-b39e-4e4f-b57f-8813951d0f1a-kube-api-access-8jxv7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2qlb4\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:56 crc kubenswrapper[4690]: I1203 13:50:56.845545 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:50:57 crc kubenswrapper[4690]: I1203 13:50:57.383054 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4"] Dec 03 13:50:57 crc kubenswrapper[4690]: I1203 13:50:57.464183 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" event={"ID":"80400476-b39e-4e4f-b57f-8813951d0f1a","Type":"ContainerStarted","Data":"92adc58714beff21acfef5e0424dae7b626d7b3d893f4b9af015507ca7e56896"} Dec 03 13:51:00 crc kubenswrapper[4690]: I1203 13:51:00.522554 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" event={"ID":"80400476-b39e-4e4f-b57f-8813951d0f1a","Type":"ContainerStarted","Data":"93840c3c1a9112b61bfed2b3f8b741c387cba82e4f683a10f1e643695a803a27"} Dec 03 13:51:00 crc kubenswrapper[4690]: I1203 13:51:00.553591 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" podStartSLOduration=2.844448534 podStartE2EDuration="4.553555065s" podCreationTimestamp="2025-12-03 13:50:56 +0000 UTC" firstStartedPulling="2025-12-03 13:50:57.408643974 +0000 UTC m=+2503.389564407" lastFinishedPulling="2025-12-03 13:50:59.117750505 +0000 UTC m=+2505.098670938" observedRunningTime="2025-12-03 13:51:00.543718788 +0000 UTC m=+2506.524639231" watchObservedRunningTime="2025-12-03 13:51:00.553555065 +0000 UTC m=+2506.534475498" Dec 03 13:51:06 crc kubenswrapper[4690]: I1203 13:51:06.315567 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:51:06 crc kubenswrapper[4690]: E1203 13:51:06.316585 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:51:07 crc kubenswrapper[4690]: I1203 13:51:07.616917 4690 generic.go:334] "Generic (PLEG): container finished" podID="80400476-b39e-4e4f-b57f-8813951d0f1a" containerID="93840c3c1a9112b61bfed2b3f8b741c387cba82e4f683a10f1e643695a803a27" exitCode=0 Dec 03 13:51:07 crc kubenswrapper[4690]: I1203 13:51:07.617006 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" event={"ID":"80400476-b39e-4e4f-b57f-8813951d0f1a","Type":"ContainerDied","Data":"93840c3c1a9112b61bfed2b3f8b741c387cba82e4f683a10f1e643695a803a27"} Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.081850 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.191300 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jxv7\" (UniqueName: \"kubernetes.io/projected/80400476-b39e-4e4f-b57f-8813951d0f1a-kube-api-access-8jxv7\") pod \"80400476-b39e-4e4f-b57f-8813951d0f1a\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.191381 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-inventory\") pod \"80400476-b39e-4e4f-b57f-8813951d0f1a\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.191548 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-ssh-key\") pod \"80400476-b39e-4e4f-b57f-8813951d0f1a\" (UID: \"80400476-b39e-4e4f-b57f-8813951d0f1a\") " Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.198172 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80400476-b39e-4e4f-b57f-8813951d0f1a-kube-api-access-8jxv7" (OuterVolumeSpecName: "kube-api-access-8jxv7") pod "80400476-b39e-4e4f-b57f-8813951d0f1a" (UID: "80400476-b39e-4e4f-b57f-8813951d0f1a"). InnerVolumeSpecName "kube-api-access-8jxv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.227041 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-inventory" (OuterVolumeSpecName: "inventory") pod "80400476-b39e-4e4f-b57f-8813951d0f1a" (UID: "80400476-b39e-4e4f-b57f-8813951d0f1a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.227343 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80400476-b39e-4e4f-b57f-8813951d0f1a" (UID: "80400476-b39e-4e4f-b57f-8813951d0f1a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.295176 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jxv7\" (UniqueName: \"kubernetes.io/projected/80400476-b39e-4e4f-b57f-8813951d0f1a-kube-api-access-8jxv7\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.295385 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.295476 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80400476-b39e-4e4f-b57f-8813951d0f1a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.638770 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" event={"ID":"80400476-b39e-4e4f-b57f-8813951d0f1a","Type":"ContainerDied","Data":"92adc58714beff21acfef5e0424dae7b626d7b3d893f4b9af015507ca7e56896"} Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.639334 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92adc58714beff21acfef5e0424dae7b626d7b3d893f4b9af015507ca7e56896" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.638851 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2qlb4" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.720847 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt"] Dec 03 13:51:09 crc kubenswrapper[4690]: E1203 13:51:09.721511 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80400476-b39e-4e4f-b57f-8813951d0f1a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.721538 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="80400476-b39e-4e4f-b57f-8813951d0f1a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.721809 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="80400476-b39e-4e4f-b57f-8813951d0f1a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.722687 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.725732 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.725891 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.725982 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.726161 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.738438 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt"] Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.808688 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj9r5\" (UniqueName: \"kubernetes.io/projected/0d10ad8c-5b1c-45c2-890a-eb42c648d437-kube-api-access-dj9r5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.809110 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.809242 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.912122 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj9r5\" (UniqueName: \"kubernetes.io/projected/0d10ad8c-5b1c-45c2-890a-eb42c648d437-kube-api-access-dj9r5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.912619 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.912739 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.918608 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.927677 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:09 crc kubenswrapper[4690]: I1203 13:51:09.933610 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj9r5\" (UniqueName: \"kubernetes.io/projected/0d10ad8c-5b1c-45c2-890a-eb42c648d437-kube-api-access-dj9r5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:10 crc kubenswrapper[4690]: I1203 13:51:10.053959 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:10 crc kubenswrapper[4690]: I1203 13:51:10.576197 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt"] Dec 03 13:51:10 crc kubenswrapper[4690]: I1203 13:51:10.649376 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" event={"ID":"0d10ad8c-5b1c-45c2-890a-eb42c648d437","Type":"ContainerStarted","Data":"eddb2134fcd03982a87c0e74b81e7a8cba4248d0122f7453a9bd7e37874b0fd9"} Dec 03 13:51:11 crc kubenswrapper[4690]: I1203 13:51:11.662510 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" event={"ID":"0d10ad8c-5b1c-45c2-890a-eb42c648d437","Type":"ContainerStarted","Data":"4ac2fc2e500f4dc8fd28b586bbd580e0bd238c6c56d32f381d47e06225a26075"} Dec 03 13:51:11 crc kubenswrapper[4690]: I1203 13:51:11.679725 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" podStartSLOduration=2.226455086 podStartE2EDuration="2.679703365s" podCreationTimestamp="2025-12-03 13:51:09 +0000 UTC" firstStartedPulling="2025-12-03 13:51:10.605315201 +0000 UTC m=+2516.586235634" lastFinishedPulling="2025-12-03 13:51:11.05856348 +0000 UTC m=+2517.039483913" observedRunningTime="2025-12-03 13:51:11.678413643 +0000 UTC m=+2517.659334076" watchObservedRunningTime="2025-12-03 13:51:11.679703365 +0000 UTC m=+2517.660623798" Dec 03 13:51:18 crc kubenswrapper[4690]: I1203 13:51:18.315125 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:51:18 crc kubenswrapper[4690]: E1203 13:51:18.316239 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:51:21 crc kubenswrapper[4690]: I1203 13:51:21.756442 4690 generic.go:334] "Generic (PLEG): container finished" podID="0d10ad8c-5b1c-45c2-890a-eb42c648d437" containerID="4ac2fc2e500f4dc8fd28b586bbd580e0bd238c6c56d32f381d47e06225a26075" exitCode=0 Dec 03 13:51:21 crc kubenswrapper[4690]: I1203 13:51:21.756840 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" event={"ID":"0d10ad8c-5b1c-45c2-890a-eb42c648d437","Type":"ContainerDied","Data":"4ac2fc2e500f4dc8fd28b586bbd580e0bd238c6c56d32f381d47e06225a26075"} Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.234694 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.320167 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-inventory\") pod \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.320760 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj9r5\" (UniqueName: \"kubernetes.io/projected/0d10ad8c-5b1c-45c2-890a-eb42c648d437-kube-api-access-dj9r5\") pod \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.321754 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-ssh-key\") pod \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\" (UID: \"0d10ad8c-5b1c-45c2-890a-eb42c648d437\") " Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.327546 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d10ad8c-5b1c-45c2-890a-eb42c648d437-kube-api-access-dj9r5" (OuterVolumeSpecName: "kube-api-access-dj9r5") pod "0d10ad8c-5b1c-45c2-890a-eb42c648d437" (UID: "0d10ad8c-5b1c-45c2-890a-eb42c648d437"). InnerVolumeSpecName "kube-api-access-dj9r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.353216 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0d10ad8c-5b1c-45c2-890a-eb42c648d437" (UID: "0d10ad8c-5b1c-45c2-890a-eb42c648d437"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.354466 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-inventory" (OuterVolumeSpecName: "inventory") pod "0d10ad8c-5b1c-45c2-890a-eb42c648d437" (UID: "0d10ad8c-5b1c-45c2-890a-eb42c648d437"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.424995 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.425036 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj9r5\" (UniqueName: \"kubernetes.io/projected/0d10ad8c-5b1c-45c2-890a-eb42c648d437-kube-api-access-dj9r5\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.425049 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d10ad8c-5b1c-45c2-890a-eb42c648d437-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.778561 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" event={"ID":"0d10ad8c-5b1c-45c2-890a-eb42c648d437","Type":"ContainerDied","Data":"eddb2134fcd03982a87c0e74b81e7a8cba4248d0122f7453a9bd7e37874b0fd9"} Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.778606 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eddb2134fcd03982a87c0e74b81e7a8cba4248d0122f7453a9bd7e37874b0fd9" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.778608 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.863988 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr"] Dec 03 13:51:23 crc kubenswrapper[4690]: E1203 13:51:23.864992 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d10ad8c-5b1c-45c2-890a-eb42c648d437" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.865096 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d10ad8c-5b1c-45c2-890a-eb42c648d437" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.865382 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d10ad8c-5b1c-45c2-890a-eb42c648d437" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.866356 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.869541 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.869777 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.870037 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.870086 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.870156 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.870467 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.870484 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.871298 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.875118 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr"] Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.937506 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb7rm\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-kube-api-access-pb7rm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.937852 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.938030 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.938143 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.938272 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.938413 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.938574 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.938711 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.938857 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.938998 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.939133 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.939321 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.939495 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:23 crc kubenswrapper[4690]: I1203 13:51:23.939535 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042009 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042105 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042137 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042176 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042210 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb7rm\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-kube-api-access-pb7rm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042237 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042277 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042307 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042362 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042428 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042479 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042523 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042577 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.042616 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.050064 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.053614 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.053773 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.060068 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.061225 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.061313 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.061331 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.061834 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.062331 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.062592 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb7rm\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-kube-api-access-pb7rm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.063072 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.063443 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.065018 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.067657 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.184579 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.759897 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr"] Dec 03 13:51:24 crc kubenswrapper[4690]: I1203 13:51:24.789908 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" event={"ID":"b529c361-6189-47c6-9c62-ea7fc0e8e30f","Type":"ContainerStarted","Data":"560baa22047dbefdf84cbbd806a11e29c801c75bbf0cde72e3aac4e01a7a34af"} Dec 03 13:51:25 crc kubenswrapper[4690]: I1203 13:51:25.801082 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" event={"ID":"b529c361-6189-47c6-9c62-ea7fc0e8e30f","Type":"ContainerStarted","Data":"fceb6cefe9115fa38e4c54b4892fd573e260e7744f184ce24da10718f4c2d877"} Dec 03 13:51:25 crc kubenswrapper[4690]: I1203 13:51:25.827159 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" podStartSLOduration=2.280971398 podStartE2EDuration="2.82713812s" podCreationTimestamp="2025-12-03 13:51:23 +0000 UTC" firstStartedPulling="2025-12-03 13:51:24.762473584 +0000 UTC m=+2530.743394017" lastFinishedPulling="2025-12-03 13:51:25.308640306 +0000 UTC m=+2531.289560739" observedRunningTime="2025-12-03 13:51:25.825747135 +0000 UTC m=+2531.806667568" watchObservedRunningTime="2025-12-03 13:51:25.82713812 +0000 UTC m=+2531.808058553" Dec 03 13:51:30 crc kubenswrapper[4690]: I1203 13:51:30.314971 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:51:30 crc kubenswrapper[4690]: E1203 13:51:30.316000 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:51:43 crc kubenswrapper[4690]: I1203 13:51:43.315963 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:51:43 crc kubenswrapper[4690]: E1203 13:51:43.317342 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:51:56 crc kubenswrapper[4690]: I1203 13:51:56.315493 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:51:56 crc kubenswrapper[4690]: E1203 13:51:56.318062 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:52:01 crc kubenswrapper[4690]: E1203 13:52:01.815456 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb529c361_6189_47c6_9c62_ea7fc0e8e30f.slice/crio-fceb6cefe9115fa38e4c54b4892fd573e260e7744f184ce24da10718f4c2d877.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb529c361_6189_47c6_9c62_ea7fc0e8e30f.slice/crio-conmon-fceb6cefe9115fa38e4c54b4892fd573e260e7744f184ce24da10718f4c2d877.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:52:02 crc kubenswrapper[4690]: I1203 13:52:02.194078 4690 generic.go:334] "Generic (PLEG): container finished" podID="b529c361-6189-47c6-9c62-ea7fc0e8e30f" containerID="fceb6cefe9115fa38e4c54b4892fd573e260e7744f184ce24da10718f4c2d877" exitCode=0 Dec 03 13:52:02 crc kubenswrapper[4690]: I1203 13:52:02.194155 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" event={"ID":"b529c361-6189-47c6-9c62-ea7fc0e8e30f","Type":"ContainerDied","Data":"fceb6cefe9115fa38e4c54b4892fd573e260e7744f184ce24da10718f4c2d877"} Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.651519 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776040 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-inventory\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776129 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776155 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-repo-setup-combined-ca-bundle\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776207 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-nova-combined-ca-bundle\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776280 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776372 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-libvirt-combined-ca-bundle\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776393 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776439 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-telemetry-combined-ca-bundle\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776488 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-bootstrap-combined-ca-bundle\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776523 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb7rm\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-kube-api-access-pb7rm\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776553 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ssh-key\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776583 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ovn-combined-ca-bundle\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776633 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.776681 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-neutron-metadata-combined-ca-bundle\") pod \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\" (UID: \"b529c361-6189-47c6-9c62-ea7fc0e8e30f\") " Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.783834 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.784543 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-kube-api-access-pb7rm" (OuterVolumeSpecName: "kube-api-access-pb7rm") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "kube-api-access-pb7rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.785564 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.786210 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.787343 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.787501 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.787586 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.787817 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.789880 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.790849 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.791059 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.796063 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.813079 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.817651 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-inventory" (OuterVolumeSpecName: "inventory") pod "b529c361-6189-47c6-9c62-ea7fc0e8e30f" (UID: "b529c361-6189-47c6-9c62-ea7fc0e8e30f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880813 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb7rm\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-kube-api-access-pb7rm\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880882 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880894 4690 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880907 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880925 4690 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880935 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880944 4690 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880955 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880967 4690 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880976 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880986 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b529c361-6189-47c6-9c62-ea7fc0e8e30f-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.880995 4690 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.881003 4690 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:03 crc kubenswrapper[4690]: I1203 13:52:03.881015 4690 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b529c361-6189-47c6-9c62-ea7fc0e8e30f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.214646 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" event={"ID":"b529c361-6189-47c6-9c62-ea7fc0e8e30f","Type":"ContainerDied","Data":"560baa22047dbefdf84cbbd806a11e29c801c75bbf0cde72e3aac4e01a7a34af"} Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.214980 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="560baa22047dbefdf84cbbd806a11e29c801c75bbf0cde72e3aac4e01a7a34af" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.214704 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.381792 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk"] Dec 03 13:52:04 crc kubenswrapper[4690]: E1203 13:52:04.382368 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b529c361-6189-47c6-9c62-ea7fc0e8e30f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.382392 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b529c361-6189-47c6-9c62-ea7fc0e8e30f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.382629 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b529c361-6189-47c6-9c62-ea7fc0e8e30f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.383398 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.394815 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.397543 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.397691 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.399742 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.400197 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.425669 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk"] Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.494692 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.494736 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.494965 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.495026 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg7n4\" (UniqueName: \"kubernetes.io/projected/2ba8f301-845c-4425-9a12-72babfccd551-kube-api-access-jg7n4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.495047 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ba8f301-845c-4425-9a12-72babfccd551-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.597239 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.597332 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg7n4\" (UniqueName: \"kubernetes.io/projected/2ba8f301-845c-4425-9a12-72babfccd551-kube-api-access-jg7n4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.597361 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ba8f301-845c-4425-9a12-72babfccd551-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.597444 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.597469 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.599093 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ba8f301-845c-4425-9a12-72babfccd551-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.603099 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.603902 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.604045 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.617060 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg7n4\" (UniqueName: \"kubernetes.io/projected/2ba8f301-845c-4425-9a12-72babfccd551-kube-api-access-jg7n4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6pqsk\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:04 crc kubenswrapper[4690]: I1203 13:52:04.710501 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:52:05 crc kubenswrapper[4690]: I1203 13:52:05.229111 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk"] Dec 03 13:52:05 crc kubenswrapper[4690]: W1203 13:52:05.235076 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ba8f301_845c_4425_9a12_72babfccd551.slice/crio-0da7a773eb3013576e92b5d5b82b9e0ef78bf451d6baa3d7e545f961b03d1702 WatchSource:0}: Error finding container 0da7a773eb3013576e92b5d5b82b9e0ef78bf451d6baa3d7e545f961b03d1702: Status 404 returned error can't find the container with id 0da7a773eb3013576e92b5d5b82b9e0ef78bf451d6baa3d7e545f961b03d1702 Dec 03 13:52:06 crc kubenswrapper[4690]: I1203 13:52:06.247441 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" event={"ID":"2ba8f301-845c-4425-9a12-72babfccd551","Type":"ContainerStarted","Data":"f713a71ed1e901a3b04ea1ffc8cd8a14c53a5098bb1bf4f71dc5537b03fc44fa"} Dec 03 13:52:06 crc kubenswrapper[4690]: I1203 13:52:06.247834 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" event={"ID":"2ba8f301-845c-4425-9a12-72babfccd551","Type":"ContainerStarted","Data":"0da7a773eb3013576e92b5d5b82b9e0ef78bf451d6baa3d7e545f961b03d1702"} Dec 03 13:52:06 crc kubenswrapper[4690]: I1203 13:52:06.271134 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" podStartSLOduration=1.789004728 podStartE2EDuration="2.271113716s" podCreationTimestamp="2025-12-03 13:52:04 +0000 UTC" firstStartedPulling="2025-12-03 13:52:05.238800465 +0000 UTC m=+2571.219720898" lastFinishedPulling="2025-12-03 13:52:05.720909453 +0000 UTC m=+2571.701829886" observedRunningTime="2025-12-03 13:52:06.262639372 +0000 UTC m=+2572.243559815" watchObservedRunningTime="2025-12-03 13:52:06.271113716 +0000 UTC m=+2572.252034149" Dec 03 13:52:07 crc kubenswrapper[4690]: I1203 13:52:07.314261 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:52:07 crc kubenswrapper[4690]: E1203 13:52:07.315013 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:52:19 crc kubenswrapper[4690]: I1203 13:52:19.315036 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:52:19 crc kubenswrapper[4690]: E1203 13:52:19.315984 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:52:31 crc kubenswrapper[4690]: I1203 13:52:31.315917 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:52:31 crc kubenswrapper[4690]: E1203 13:52:31.316889 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:52:42 crc kubenswrapper[4690]: I1203 13:52:42.315192 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:52:42 crc kubenswrapper[4690]: E1203 13:52:42.316287 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:52:57 crc kubenswrapper[4690]: I1203 13:52:57.315438 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:52:57 crc kubenswrapper[4690]: E1203 13:52:57.316288 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:53:08 crc kubenswrapper[4690]: I1203 13:53:08.315802 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:53:08 crc kubenswrapper[4690]: E1203 13:53:08.316695 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:53:12 crc kubenswrapper[4690]: I1203 13:53:12.958380 4690 generic.go:334] "Generic (PLEG): container finished" podID="2ba8f301-845c-4425-9a12-72babfccd551" containerID="f713a71ed1e901a3b04ea1ffc8cd8a14c53a5098bb1bf4f71dc5537b03fc44fa" exitCode=0 Dec 03 13:53:12 crc kubenswrapper[4690]: I1203 13:53:12.958456 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" event={"ID":"2ba8f301-845c-4425-9a12-72babfccd551","Type":"ContainerDied","Data":"f713a71ed1e901a3b04ea1ffc8cd8a14c53a5098bb1bf4f71dc5537b03fc44fa"} Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.419274 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.553536 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg7n4\" (UniqueName: \"kubernetes.io/projected/2ba8f301-845c-4425-9a12-72babfccd551-kube-api-access-jg7n4\") pod \"2ba8f301-845c-4425-9a12-72babfccd551\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.553596 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ssh-key\") pod \"2ba8f301-845c-4425-9a12-72babfccd551\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.553800 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ovn-combined-ca-bundle\") pod \"2ba8f301-845c-4425-9a12-72babfccd551\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.553858 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-inventory\") pod \"2ba8f301-845c-4425-9a12-72babfccd551\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.553923 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ba8f301-845c-4425-9a12-72babfccd551-ovncontroller-config-0\") pod \"2ba8f301-845c-4425-9a12-72babfccd551\" (UID: \"2ba8f301-845c-4425-9a12-72babfccd551\") " Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.560116 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2ba8f301-845c-4425-9a12-72babfccd551" (UID: "2ba8f301-845c-4425-9a12-72babfccd551"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.560173 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ba8f301-845c-4425-9a12-72babfccd551-kube-api-access-jg7n4" (OuterVolumeSpecName: "kube-api-access-jg7n4") pod "2ba8f301-845c-4425-9a12-72babfccd551" (UID: "2ba8f301-845c-4425-9a12-72babfccd551"). InnerVolumeSpecName "kube-api-access-jg7n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.585313 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ba8f301-845c-4425-9a12-72babfccd551" (UID: "2ba8f301-845c-4425-9a12-72babfccd551"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.593901 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ba8f301-845c-4425-9a12-72babfccd551-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "2ba8f301-845c-4425-9a12-72babfccd551" (UID: "2ba8f301-845c-4425-9a12-72babfccd551"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.601067 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-inventory" (OuterVolumeSpecName: "inventory") pod "2ba8f301-845c-4425-9a12-72babfccd551" (UID: "2ba8f301-845c-4425-9a12-72babfccd551"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.656622 4690 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.656977 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.656991 4690 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ba8f301-845c-4425-9a12-72babfccd551-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.657001 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg7n4\" (UniqueName: \"kubernetes.io/projected/2ba8f301-845c-4425-9a12-72babfccd551-kube-api-access-jg7n4\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.657012 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba8f301-845c-4425-9a12-72babfccd551-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.982830 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" event={"ID":"2ba8f301-845c-4425-9a12-72babfccd551","Type":"ContainerDied","Data":"0da7a773eb3013576e92b5d5b82b9e0ef78bf451d6baa3d7e545f961b03d1702"} Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.982903 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0da7a773eb3013576e92b5d5b82b9e0ef78bf451d6baa3d7e545f961b03d1702" Dec 03 13:53:14 crc kubenswrapper[4690]: I1203 13:53:14.982909 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6pqsk" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.063796 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz"] Dec 03 13:53:15 crc kubenswrapper[4690]: E1203 13:53:15.064353 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba8f301-845c-4425-9a12-72babfccd551" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.064374 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba8f301-845c-4425-9a12-72babfccd551" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.064598 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba8f301-845c-4425-9a12-72babfccd551" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.065390 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.069513 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.069968 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.070162 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.070424 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.070576 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.074851 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.076946 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz"] Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.168471 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.168688 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.168850 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zrfr\" (UniqueName: \"kubernetes.io/projected/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-kube-api-access-2zrfr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.168921 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.168980 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.169082 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.271477 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.271556 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.271604 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.271736 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.271762 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.271806 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zrfr\" (UniqueName: \"kubernetes.io/projected/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-kube-api-access-2zrfr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.279351 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.279388 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.279886 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.280334 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.307747 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zrfr\" (UniqueName: \"kubernetes.io/projected/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-kube-api-access-2zrfr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.317338 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.390554 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.967081 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz"] Dec 03 13:53:15 crc kubenswrapper[4690]: I1203 13:53:15.995740 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" event={"ID":"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8","Type":"ContainerStarted","Data":"5ca3f09290d14aac8d4d028d71f2dd02417b5b74b3737a2a5430bfbf12fd1b24"} Dec 03 13:53:18 crc kubenswrapper[4690]: I1203 13:53:18.016829 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" event={"ID":"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8","Type":"ContainerStarted","Data":"99ac53bc9b2346567aa91e049441182b8edb7e07810d121ba6d3458eff2e25bb"} Dec 03 13:53:19 crc kubenswrapper[4690]: I1203 13:53:19.315301 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:53:19 crc kubenswrapper[4690]: E1203 13:53:19.316713 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:53:31 crc kubenswrapper[4690]: I1203 13:53:31.315179 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:53:31 crc kubenswrapper[4690]: E1203 13:53:31.316330 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:53:43 crc kubenswrapper[4690]: I1203 13:53:43.315371 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:53:43 crc kubenswrapper[4690]: E1203 13:53:43.316391 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:53:58 crc kubenswrapper[4690]: I1203 13:53:58.315590 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:53:58 crc kubenswrapper[4690]: E1203 13:53:58.316567 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:54:07 crc kubenswrapper[4690]: I1203 13:54:07.937175 4690 generic.go:334] "Generic (PLEG): container finished" podID="b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" containerID="99ac53bc9b2346567aa91e049441182b8edb7e07810d121ba6d3458eff2e25bb" exitCode=0 Dec 03 13:54:07 crc kubenswrapper[4690]: I1203 13:54:07.937275 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" event={"ID":"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8","Type":"ContainerDied","Data":"99ac53bc9b2346567aa91e049441182b8edb7e07810d121ba6d3458eff2e25bb"} Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.441039 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.506083 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.506144 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-metadata-combined-ca-bundle\") pod \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.506236 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-inventory\") pod \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.506351 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-ssh-key\") pod \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.506522 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zrfr\" (UniqueName: \"kubernetes.io/projected/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-kube-api-access-2zrfr\") pod \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.506578 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-nova-metadata-neutron-config-0\") pod \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\" (UID: \"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8\") " Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.515043 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" (UID: "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.517066 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-kube-api-access-2zrfr" (OuterVolumeSpecName: "kube-api-access-2zrfr") pod "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" (UID: "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8"). InnerVolumeSpecName "kube-api-access-2zrfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.540227 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" (UID: "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.541331 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" (UID: "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.542610 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" (UID: "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.554339 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-inventory" (OuterVolumeSpecName: "inventory") pod "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" (UID: "b1d5dc5c-d0a9-4194-88df-b48e39cad7d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.609206 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zrfr\" (UniqueName: \"kubernetes.io/projected/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-kube-api-access-2zrfr\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.609261 4690 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.609278 4690 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.609293 4690 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.609305 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.609314 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1d5dc5c-d0a9-4194-88df-b48e39cad7d8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.961994 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" event={"ID":"b1d5dc5c-d0a9-4194-88df-b48e39cad7d8","Type":"ContainerDied","Data":"5ca3f09290d14aac8d4d028d71f2dd02417b5b74b3737a2a5430bfbf12fd1b24"} Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.962043 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ca3f09290d14aac8d4d028d71f2dd02417b5b74b3737a2a5430bfbf12fd1b24" Dec 03 13:54:09 crc kubenswrapper[4690]: I1203 13:54:09.962046 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.074282 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq"] Dec 03 13:54:10 crc kubenswrapper[4690]: E1203 13:54:10.074764 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.074781 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.075000 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d5dc5c-d0a9-4194-88df-b48e39cad7d8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.075697 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.084156 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.084171 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.084230 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.084451 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.084536 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.108376 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq"] Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.223382 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.223447 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blv9b\" (UniqueName: \"kubernetes.io/projected/8ce19539-f48f-446b-8a79-d528d58a9e8d-kube-api-access-blv9b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.223494 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.223585 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.224005 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.325938 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.326023 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.326060 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blv9b\" (UniqueName: \"kubernetes.io/projected/8ce19539-f48f-446b-8a79-d528d58a9e8d-kube-api-access-blv9b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.326116 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.326195 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.329861 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.330226 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.330497 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.330792 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.345629 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blv9b\" (UniqueName: \"kubernetes.io/projected/8ce19539-f48f-446b-8a79-d528d58a9e8d-kube-api-access-blv9b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.396972 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.921935 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq"] Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.932249 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:54:10 crc kubenswrapper[4690]: I1203 13:54:10.974011 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" event={"ID":"8ce19539-f48f-446b-8a79-d528d58a9e8d","Type":"ContainerStarted","Data":"728d43846cdc19398445756d1892408a23b8ded277d67fe389a997a6338214d2"} Dec 03 13:54:11 crc kubenswrapper[4690]: I1203 13:54:11.314910 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:54:11 crc kubenswrapper[4690]: E1203 13:54:11.315156 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 13:54:11 crc kubenswrapper[4690]: I1203 13:54:11.987986 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" event={"ID":"8ce19539-f48f-446b-8a79-d528d58a9e8d","Type":"ContainerStarted","Data":"710c5ac2244240fa65af3a200849cc635f34e19683acf9b68229e8d91ac2587a"} Dec 03 13:54:12 crc kubenswrapper[4690]: I1203 13:54:12.009098 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" podStartSLOduration=1.355554315 podStartE2EDuration="2.009071804s" podCreationTimestamp="2025-12-03 13:54:10 +0000 UTC" firstStartedPulling="2025-12-03 13:54:10.932003808 +0000 UTC m=+2696.912924241" lastFinishedPulling="2025-12-03 13:54:11.585521297 +0000 UTC m=+2697.566441730" observedRunningTime="2025-12-03 13:54:12.003121124 +0000 UTC m=+2697.984041567" watchObservedRunningTime="2025-12-03 13:54:12.009071804 +0000 UTC m=+2697.989992227" Dec 03 13:54:23 crc kubenswrapper[4690]: I1203 13:54:23.314805 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:54:24 crc kubenswrapper[4690]: I1203 13:54:24.106566 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"f2186a489870b5492e9f2c87bb858a3f5a82bbdd59d220ed13a4486d7275a008"} Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.761657 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6gt95"] Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.765275 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.773024 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6gt95"] Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.782219 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-catalog-content\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.782300 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntrv2\" (UniqueName: \"kubernetes.io/projected/02f8d43d-3328-4bd3-a36c-3c1e4e892016-kube-api-access-ntrv2\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.782580 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-utilities\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.884661 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-utilities\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.885146 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-catalog-content\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.885271 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntrv2\" (UniqueName: \"kubernetes.io/projected/02f8d43d-3328-4bd3-a36c-3c1e4e892016-kube-api-access-ntrv2\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.885366 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-utilities\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.885586 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-catalog-content\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:43 crc kubenswrapper[4690]: I1203 13:55:43.917025 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntrv2\" (UniqueName: \"kubernetes.io/projected/02f8d43d-3328-4bd3-a36c-3c1e4e892016-kube-api-access-ntrv2\") pod \"community-operators-6gt95\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:44 crc kubenswrapper[4690]: I1203 13:55:44.091488 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:44 crc kubenswrapper[4690]: I1203 13:55:44.685743 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6gt95"] Dec 03 13:55:44 crc kubenswrapper[4690]: I1203 13:55:44.924560 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gt95" event={"ID":"02f8d43d-3328-4bd3-a36c-3c1e4e892016","Type":"ContainerStarted","Data":"8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb"} Dec 03 13:55:44 crc kubenswrapper[4690]: I1203 13:55:44.924942 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gt95" event={"ID":"02f8d43d-3328-4bd3-a36c-3c1e4e892016","Type":"ContainerStarted","Data":"896c98de31bb4fb876a9f042152ef2b15dc59af52927ee17b0148c40efb46dee"} Dec 03 13:55:45 crc kubenswrapper[4690]: I1203 13:55:45.937784 4690 generic.go:334] "Generic (PLEG): container finished" podID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerID="8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb" exitCode=0 Dec 03 13:55:45 crc kubenswrapper[4690]: I1203 13:55:45.937849 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gt95" event={"ID":"02f8d43d-3328-4bd3-a36c-3c1e4e892016","Type":"ContainerDied","Data":"8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb"} Dec 03 13:55:47 crc kubenswrapper[4690]: I1203 13:55:47.964414 4690 generic.go:334] "Generic (PLEG): container finished" podID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerID="38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360" exitCode=0 Dec 03 13:55:47 crc kubenswrapper[4690]: I1203 13:55:47.965018 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gt95" event={"ID":"02f8d43d-3328-4bd3-a36c-3c1e4e892016","Type":"ContainerDied","Data":"38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360"} Dec 03 13:55:48 crc kubenswrapper[4690]: I1203 13:55:48.980851 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gt95" event={"ID":"02f8d43d-3328-4bd3-a36c-3c1e4e892016","Type":"ContainerStarted","Data":"7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81"} Dec 03 13:55:49 crc kubenswrapper[4690]: I1203 13:55:49.007166 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6gt95" podStartSLOduration=3.532216521 podStartE2EDuration="6.007142705s" podCreationTimestamp="2025-12-03 13:55:43 +0000 UTC" firstStartedPulling="2025-12-03 13:55:45.940099099 +0000 UTC m=+2791.921019532" lastFinishedPulling="2025-12-03 13:55:48.415025263 +0000 UTC m=+2794.395945716" observedRunningTime="2025-12-03 13:55:48.996499907 +0000 UTC m=+2794.977420350" watchObservedRunningTime="2025-12-03 13:55:49.007142705 +0000 UTC m=+2794.988063138" Dec 03 13:55:54 crc kubenswrapper[4690]: I1203 13:55:54.092580 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:54 crc kubenswrapper[4690]: I1203 13:55:54.093374 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:54 crc kubenswrapper[4690]: I1203 13:55:54.138659 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:55 crc kubenswrapper[4690]: I1203 13:55:55.084424 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:55 crc kubenswrapper[4690]: I1203 13:55:55.133594 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6gt95"] Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.058241 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6gt95" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerName="registry-server" containerID="cri-o://7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81" gracePeriod=2 Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.518198 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.723202 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-utilities\") pod \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.723280 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-catalog-content\") pod \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.723361 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntrv2\" (UniqueName: \"kubernetes.io/projected/02f8d43d-3328-4bd3-a36c-3c1e4e892016-kube-api-access-ntrv2\") pod \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\" (UID: \"02f8d43d-3328-4bd3-a36c-3c1e4e892016\") " Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.724016 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-utilities" (OuterVolumeSpecName: "utilities") pod "02f8d43d-3328-4bd3-a36c-3c1e4e892016" (UID: "02f8d43d-3328-4bd3-a36c-3c1e4e892016"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.724522 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.733720 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f8d43d-3328-4bd3-a36c-3c1e4e892016-kube-api-access-ntrv2" (OuterVolumeSpecName: "kube-api-access-ntrv2") pod "02f8d43d-3328-4bd3-a36c-3c1e4e892016" (UID: "02f8d43d-3328-4bd3-a36c-3c1e4e892016"). InnerVolumeSpecName "kube-api-access-ntrv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.775258 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02f8d43d-3328-4bd3-a36c-3c1e4e892016" (UID: "02f8d43d-3328-4bd3-a36c-3c1e4e892016"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.826400 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f8d43d-3328-4bd3-a36c-3c1e4e892016-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:55:57 crc kubenswrapper[4690]: I1203 13:55:57.826436 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntrv2\" (UniqueName: \"kubernetes.io/projected/02f8d43d-3328-4bd3-a36c-3c1e4e892016-kube-api-access-ntrv2\") on node \"crc\" DevicePath \"\"" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.068954 4690 generic.go:334] "Generic (PLEG): container finished" podID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerID="7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81" exitCode=0 Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.068998 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gt95" event={"ID":"02f8d43d-3328-4bd3-a36c-3c1e4e892016","Type":"ContainerDied","Data":"7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81"} Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.069063 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gt95" event={"ID":"02f8d43d-3328-4bd3-a36c-3c1e4e892016","Type":"ContainerDied","Data":"896c98de31bb4fb876a9f042152ef2b15dc59af52927ee17b0148c40efb46dee"} Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.069063 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6gt95" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.069080 4690 scope.go:117] "RemoveContainer" containerID="7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.106979 4690 scope.go:117] "RemoveContainer" containerID="38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.113075 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6gt95"] Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.121995 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6gt95"] Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.128693 4690 scope.go:117] "RemoveContainer" containerID="8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.186095 4690 scope.go:117] "RemoveContainer" containerID="7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81" Dec 03 13:55:58 crc kubenswrapper[4690]: E1203 13:55:58.190053 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81\": container with ID starting with 7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81 not found: ID does not exist" containerID="7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.190126 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81"} err="failed to get container status \"7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81\": rpc error: code = NotFound desc = could not find container \"7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81\": container with ID starting with 7476c49fd92465e7db1f724dd9c73516d3702ef0b4266a7c2325c8407008dd81 not found: ID does not exist" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.190171 4690 scope.go:117] "RemoveContainer" containerID="38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360" Dec 03 13:55:58 crc kubenswrapper[4690]: E1203 13:55:58.190710 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360\": container with ID starting with 38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360 not found: ID does not exist" containerID="38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.190769 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360"} err="failed to get container status \"38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360\": rpc error: code = NotFound desc = could not find container \"38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360\": container with ID starting with 38781f2bc0a17ed4125c655046734d53288f4157f7073d9a053d0124939e0360 not found: ID does not exist" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.190808 4690 scope.go:117] "RemoveContainer" containerID="8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb" Dec 03 13:55:58 crc kubenswrapper[4690]: E1203 13:55:58.191167 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb\": container with ID starting with 8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb not found: ID does not exist" containerID="8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.191203 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb"} err="failed to get container status \"8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb\": rpc error: code = NotFound desc = could not find container \"8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb\": container with ID starting with 8f662dc6862ada46d33124403090ba10a24c63c785ce9a1c74fe04c2cfa944cb not found: ID does not exist" Dec 03 13:55:58 crc kubenswrapper[4690]: E1203 13:55:58.199880 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02f8d43d_3328_4bd3_a36c_3c1e4e892016.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02f8d43d_3328_4bd3_a36c_3c1e4e892016.slice/crio-896c98de31bb4fb876a9f042152ef2b15dc59af52927ee17b0148c40efb46dee\": RecentStats: unable to find data in memory cache]" Dec 03 13:55:58 crc kubenswrapper[4690]: I1203 13:55:58.327942 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" path="/var/lib/kubelet/pods/02f8d43d-3328-4bd3-a36c-3c1e4e892016/volumes" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.054293 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nzxbg"] Dec 03 13:56:16 crc kubenswrapper[4690]: E1203 13:56:16.056706 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerName="extract-content" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.056833 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerName="extract-content" Dec 03 13:56:16 crc kubenswrapper[4690]: E1203 13:56:16.056954 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerName="extract-utilities" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.057045 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerName="extract-utilities" Dec 03 13:56:16 crc kubenswrapper[4690]: E1203 13:56:16.057159 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerName="registry-server" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.057244 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerName="registry-server" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.057680 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f8d43d-3328-4bd3-a36c-3c1e4e892016" containerName="registry-server" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.059727 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.070833 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nzxbg"] Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.229770 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-utilities\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.229829 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmlrg\" (UniqueName: \"kubernetes.io/projected/89021641-e329-4b28-b275-f3dc73e4fa47-kube-api-access-rmlrg\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.229924 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-catalog-content\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.332482 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-utilities\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.332557 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmlrg\" (UniqueName: \"kubernetes.io/projected/89021641-e329-4b28-b275-f3dc73e4fa47-kube-api-access-rmlrg\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.332608 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-catalog-content\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.333019 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-utilities\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.333084 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-catalog-content\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.357592 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmlrg\" (UniqueName: \"kubernetes.io/projected/89021641-e329-4b28-b275-f3dc73e4fa47-kube-api-access-rmlrg\") pod \"certified-operators-nzxbg\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.381711 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:16 crc kubenswrapper[4690]: I1203 13:56:16.984341 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nzxbg"] Dec 03 13:56:17 crc kubenswrapper[4690]: I1203 13:56:17.254530 4690 generic.go:334] "Generic (PLEG): container finished" podID="89021641-e329-4b28-b275-f3dc73e4fa47" containerID="6a8851103eb3243c74e3caa696f0922d6e7aa42127a7928b6ba7b460a9c22e48" exitCode=0 Dec 03 13:56:17 crc kubenswrapper[4690]: I1203 13:56:17.254587 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzxbg" event={"ID":"89021641-e329-4b28-b275-f3dc73e4fa47","Type":"ContainerDied","Data":"6a8851103eb3243c74e3caa696f0922d6e7aa42127a7928b6ba7b460a9c22e48"} Dec 03 13:56:17 crc kubenswrapper[4690]: I1203 13:56:17.254648 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzxbg" event={"ID":"89021641-e329-4b28-b275-f3dc73e4fa47","Type":"ContainerStarted","Data":"ee330380a324b29dffe76485de975f6dbe4ec1e056a84b72f90604039202fbe1"} Dec 03 13:56:18 crc kubenswrapper[4690]: I1203 13:56:18.278320 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzxbg" event={"ID":"89021641-e329-4b28-b275-f3dc73e4fa47","Type":"ContainerStarted","Data":"978667c59166fabd7f88b0974895d9df0204c2289dd6cff40f72a568b1703b22"} Dec 03 13:56:19 crc kubenswrapper[4690]: I1203 13:56:19.312099 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzxbg" event={"ID":"89021641-e329-4b28-b275-f3dc73e4fa47","Type":"ContainerDied","Data":"978667c59166fabd7f88b0974895d9df0204c2289dd6cff40f72a568b1703b22"} Dec 03 13:56:19 crc kubenswrapper[4690]: I1203 13:56:19.312250 4690 generic.go:334] "Generic (PLEG): container finished" podID="89021641-e329-4b28-b275-f3dc73e4fa47" containerID="978667c59166fabd7f88b0974895d9df0204c2289dd6cff40f72a568b1703b22" exitCode=0 Dec 03 13:56:20 crc kubenswrapper[4690]: I1203 13:56:20.329018 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzxbg" event={"ID":"89021641-e329-4b28-b275-f3dc73e4fa47","Type":"ContainerStarted","Data":"628b6e4d1fe5abe1dd9b643c9aeb7eb7bf28259d5d5fd4469ecf806452997088"} Dec 03 13:56:20 crc kubenswrapper[4690]: I1203 13:56:20.360516 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nzxbg" podStartSLOduration=1.872965873 podStartE2EDuration="4.360490715s" podCreationTimestamp="2025-12-03 13:56:16 +0000 UTC" firstStartedPulling="2025-12-03 13:56:17.257279198 +0000 UTC m=+2823.238199631" lastFinishedPulling="2025-12-03 13:56:19.74480404 +0000 UTC m=+2825.725724473" observedRunningTime="2025-12-03 13:56:20.34796448 +0000 UTC m=+2826.328884913" watchObservedRunningTime="2025-12-03 13:56:20.360490715 +0000 UTC m=+2826.341411148" Dec 03 13:56:26 crc kubenswrapper[4690]: I1203 13:56:26.382956 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:26 crc kubenswrapper[4690]: I1203 13:56:26.383536 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:26 crc kubenswrapper[4690]: I1203 13:56:26.437215 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:27 crc kubenswrapper[4690]: I1203 13:56:27.451125 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:28 crc kubenswrapper[4690]: I1203 13:56:28.671437 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nzxbg"] Dec 03 13:56:29 crc kubenswrapper[4690]: I1203 13:56:29.426362 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nzxbg" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" containerName="registry-server" containerID="cri-o://628b6e4d1fe5abe1dd9b643c9aeb7eb7bf28259d5d5fd4469ecf806452997088" gracePeriod=2 Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.441195 4690 generic.go:334] "Generic (PLEG): container finished" podID="89021641-e329-4b28-b275-f3dc73e4fa47" containerID="628b6e4d1fe5abe1dd9b643c9aeb7eb7bf28259d5d5fd4469ecf806452997088" exitCode=0 Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.441267 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzxbg" event={"ID":"89021641-e329-4b28-b275-f3dc73e4fa47","Type":"ContainerDied","Data":"628b6e4d1fe5abe1dd9b643c9aeb7eb7bf28259d5d5fd4469ecf806452997088"} Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.441921 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzxbg" event={"ID":"89021641-e329-4b28-b275-f3dc73e4fa47","Type":"ContainerDied","Data":"ee330380a324b29dffe76485de975f6dbe4ec1e056a84b72f90604039202fbe1"} Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.441941 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee330380a324b29dffe76485de975f6dbe4ec1e056a84b72f90604039202fbe1" Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.445939 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.561532 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-utilities\") pod \"89021641-e329-4b28-b275-f3dc73e4fa47\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.561627 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-catalog-content\") pod \"89021641-e329-4b28-b275-f3dc73e4fa47\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.561687 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmlrg\" (UniqueName: \"kubernetes.io/projected/89021641-e329-4b28-b275-f3dc73e4fa47-kube-api-access-rmlrg\") pod \"89021641-e329-4b28-b275-f3dc73e4fa47\" (UID: \"89021641-e329-4b28-b275-f3dc73e4fa47\") " Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.564545 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-utilities" (OuterVolumeSpecName: "utilities") pod "89021641-e329-4b28-b275-f3dc73e4fa47" (UID: "89021641-e329-4b28-b275-f3dc73e4fa47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.570636 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89021641-e329-4b28-b275-f3dc73e4fa47-kube-api-access-rmlrg" (OuterVolumeSpecName: "kube-api-access-rmlrg") pod "89021641-e329-4b28-b275-f3dc73e4fa47" (UID: "89021641-e329-4b28-b275-f3dc73e4fa47"). InnerVolumeSpecName "kube-api-access-rmlrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.613787 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89021641-e329-4b28-b275-f3dc73e4fa47" (UID: "89021641-e329-4b28-b275-f3dc73e4fa47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.664693 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.664743 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89021641-e329-4b28-b275-f3dc73e4fa47-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:56:30 crc kubenswrapper[4690]: I1203 13:56:30.664757 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmlrg\" (UniqueName: \"kubernetes.io/projected/89021641-e329-4b28-b275-f3dc73e4fa47-kube-api-access-rmlrg\") on node \"crc\" DevicePath \"\"" Dec 03 13:56:31 crc kubenswrapper[4690]: I1203 13:56:31.451276 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzxbg" Dec 03 13:56:31 crc kubenswrapper[4690]: I1203 13:56:31.532023 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nzxbg"] Dec 03 13:56:31 crc kubenswrapper[4690]: I1203 13:56:31.545310 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nzxbg"] Dec 03 13:56:32 crc kubenswrapper[4690]: I1203 13:56:32.330231 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" path="/var/lib/kubelet/pods/89021641-e329-4b28-b275-f3dc73e4fa47/volumes" Dec 03 13:56:46 crc kubenswrapper[4690]: I1203 13:56:46.823806 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:56:46 crc kubenswrapper[4690]: I1203 13:56:46.824651 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:57:16 crc kubenswrapper[4690]: I1203 13:57:16.824059 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:57:16 crc kubenswrapper[4690]: I1203 13:57:16.824623 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:57:46 crc kubenswrapper[4690]: I1203 13:57:46.823939 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:57:46 crc kubenswrapper[4690]: I1203 13:57:46.824854 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:57:46 crc kubenswrapper[4690]: I1203 13:57:46.824939 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 13:57:46 crc kubenswrapper[4690]: I1203 13:57:46.825922 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2186a489870b5492e9f2c87bb858a3f5a82bbdd59d220ed13a4486d7275a008"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:57:46 crc kubenswrapper[4690]: I1203 13:57:46.825983 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://f2186a489870b5492e9f2c87bb858a3f5a82bbdd59d220ed13a4486d7275a008" gracePeriod=600 Dec 03 13:57:47 crc kubenswrapper[4690]: I1203 13:57:47.197691 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="f2186a489870b5492e9f2c87bb858a3f5a82bbdd59d220ed13a4486d7275a008" exitCode=0 Dec 03 13:57:47 crc kubenswrapper[4690]: I1203 13:57:47.197775 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"f2186a489870b5492e9f2c87bb858a3f5a82bbdd59d220ed13a4486d7275a008"} Dec 03 13:57:47 crc kubenswrapper[4690]: I1203 13:57:47.198276 4690 scope.go:117] "RemoveContainer" containerID="cfe6a7947977042caa103243fb94eeb1b72adf0ab9bf42ea365ac3fd1dfe13a1" Dec 03 13:57:48 crc kubenswrapper[4690]: I1203 13:57:48.212146 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615"} Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.735695 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7slb7"] Dec 03 13:58:34 crc kubenswrapper[4690]: E1203 13:58:34.738294 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" containerName="registry-server" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.738427 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" containerName="registry-server" Dec 03 13:58:34 crc kubenswrapper[4690]: E1203 13:58:34.738560 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" containerName="extract-content" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.738635 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" containerName="extract-content" Dec 03 13:58:34 crc kubenswrapper[4690]: E1203 13:58:34.738718 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" containerName="extract-utilities" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.738799 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" containerName="extract-utilities" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.739181 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="89021641-e329-4b28-b275-f3dc73e4fa47" containerName="registry-server" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.741428 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.750159 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7slb7"] Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.846017 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shngx\" (UniqueName: \"kubernetes.io/projected/1d7d24dd-562c-4ee3-856a-57f55ba404f3-kube-api-access-shngx\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.846462 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-utilities\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.846569 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-catalog-content\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.949194 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shngx\" (UniqueName: \"kubernetes.io/projected/1d7d24dd-562c-4ee3-856a-57f55ba404f3-kube-api-access-shngx\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.949349 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-utilities\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.949388 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-catalog-content\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.950114 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-catalog-content\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.950409 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-utilities\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:34 crc kubenswrapper[4690]: I1203 13:58:34.981138 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shngx\" (UniqueName: \"kubernetes.io/projected/1d7d24dd-562c-4ee3-856a-57f55ba404f3-kube-api-access-shngx\") pod \"redhat-marketplace-7slb7\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:35 crc kubenswrapper[4690]: I1203 13:58:35.082560 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:35 crc kubenswrapper[4690]: I1203 13:58:35.702916 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7slb7"] Dec 03 13:58:35 crc kubenswrapper[4690]: W1203 13:58:35.704802 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d7d24dd_562c_4ee3_856a_57f55ba404f3.slice/crio-8ea0d392a39aa9d557cb702fdeda43f56b522f8b5b807605b8b82c71dddcd25f WatchSource:0}: Error finding container 8ea0d392a39aa9d557cb702fdeda43f56b522f8b5b807605b8b82c71dddcd25f: Status 404 returned error can't find the container with id 8ea0d392a39aa9d557cb702fdeda43f56b522f8b5b807605b8b82c71dddcd25f Dec 03 13:58:36 crc kubenswrapper[4690]: I1203 13:58:36.686851 4690 generic.go:334] "Generic (PLEG): container finished" podID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerID="529df63edc24ad7931817f5233c78074c07333be77341fc965a4d8c868151b5d" exitCode=0 Dec 03 13:58:36 crc kubenswrapper[4690]: I1203 13:58:36.686996 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7slb7" event={"ID":"1d7d24dd-562c-4ee3-856a-57f55ba404f3","Type":"ContainerDied","Data":"529df63edc24ad7931817f5233c78074c07333be77341fc965a4d8c868151b5d"} Dec 03 13:58:36 crc kubenswrapper[4690]: I1203 13:58:36.687303 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7slb7" event={"ID":"1d7d24dd-562c-4ee3-856a-57f55ba404f3","Type":"ContainerStarted","Data":"8ea0d392a39aa9d557cb702fdeda43f56b522f8b5b807605b8b82c71dddcd25f"} Dec 03 13:58:37 crc kubenswrapper[4690]: I1203 13:58:37.700215 4690 generic.go:334] "Generic (PLEG): container finished" podID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerID="561e985c1e2558ece5148dadcd72fa2305050daa13dd71d85c69c5deda8ed240" exitCode=0 Dec 03 13:58:37 crc kubenswrapper[4690]: I1203 13:58:37.701176 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7slb7" event={"ID":"1d7d24dd-562c-4ee3-856a-57f55ba404f3","Type":"ContainerDied","Data":"561e985c1e2558ece5148dadcd72fa2305050daa13dd71d85c69c5deda8ed240"} Dec 03 13:58:37 crc kubenswrapper[4690]: I1203 13:58:37.704195 4690 generic.go:334] "Generic (PLEG): container finished" podID="8ce19539-f48f-446b-8a79-d528d58a9e8d" containerID="710c5ac2244240fa65af3a200849cc635f34e19683acf9b68229e8d91ac2587a" exitCode=0 Dec 03 13:58:37 crc kubenswrapper[4690]: I1203 13:58:37.704233 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" event={"ID":"8ce19539-f48f-446b-8a79-d528d58a9e8d","Type":"ContainerDied","Data":"710c5ac2244240fa65af3a200849cc635f34e19683acf9b68229e8d91ac2587a"} Dec 03 13:58:38 crc kubenswrapper[4690]: I1203 13:58:38.717139 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7slb7" event={"ID":"1d7d24dd-562c-4ee3-856a-57f55ba404f3","Type":"ContainerStarted","Data":"02e955637cb0c6c5fd9781f4852fed62aed8c6b372c826d5e249215b898e5b41"} Dec 03 13:58:38 crc kubenswrapper[4690]: I1203 13:58:38.735306 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7slb7" podStartSLOduration=3.249784269 podStartE2EDuration="4.735280946s" podCreationTimestamp="2025-12-03 13:58:34 +0000 UTC" firstStartedPulling="2025-12-03 13:58:36.688963944 +0000 UTC m=+2962.669884377" lastFinishedPulling="2025-12-03 13:58:38.174460621 +0000 UTC m=+2964.155381054" observedRunningTime="2025-12-03 13:58:38.733943213 +0000 UTC m=+2964.714863706" watchObservedRunningTime="2025-12-03 13:58:38.735280946 +0000 UTC m=+2964.716201379" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.228536 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.360210 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-secret-0\") pod \"8ce19539-f48f-446b-8a79-d528d58a9e8d\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.360324 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-inventory\") pod \"8ce19539-f48f-446b-8a79-d528d58a9e8d\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.360389 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blv9b\" (UniqueName: \"kubernetes.io/projected/8ce19539-f48f-446b-8a79-d528d58a9e8d-kube-api-access-blv9b\") pod \"8ce19539-f48f-446b-8a79-d528d58a9e8d\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.360407 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-ssh-key\") pod \"8ce19539-f48f-446b-8a79-d528d58a9e8d\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.360507 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-combined-ca-bundle\") pod \"8ce19539-f48f-446b-8a79-d528d58a9e8d\" (UID: \"8ce19539-f48f-446b-8a79-d528d58a9e8d\") " Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.369260 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8ce19539-f48f-446b-8a79-d528d58a9e8d" (UID: "8ce19539-f48f-446b-8a79-d528d58a9e8d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.369467 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ce19539-f48f-446b-8a79-d528d58a9e8d-kube-api-access-blv9b" (OuterVolumeSpecName: "kube-api-access-blv9b") pod "8ce19539-f48f-446b-8a79-d528d58a9e8d" (UID: "8ce19539-f48f-446b-8a79-d528d58a9e8d"). InnerVolumeSpecName "kube-api-access-blv9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.390160 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "8ce19539-f48f-446b-8a79-d528d58a9e8d" (UID: "8ce19539-f48f-446b-8a79-d528d58a9e8d"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.398007 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-inventory" (OuterVolumeSpecName: "inventory") pod "8ce19539-f48f-446b-8a79-d528d58a9e8d" (UID: "8ce19539-f48f-446b-8a79-d528d58a9e8d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.414117 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8ce19539-f48f-446b-8a79-d528d58a9e8d" (UID: "8ce19539-f48f-446b-8a79-d528d58a9e8d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.463529 4690 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.463774 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.463882 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.464008 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blv9b\" (UniqueName: \"kubernetes.io/projected/8ce19539-f48f-446b-8a79-d528d58a9e8d-kube-api-access-blv9b\") on node \"crc\" DevicePath \"\"" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.464116 4690 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce19539-f48f-446b-8a79-d528d58a9e8d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.727520 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.727521 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq" event={"ID":"8ce19539-f48f-446b-8a79-d528d58a9e8d","Type":"ContainerDied","Data":"728d43846cdc19398445756d1892408a23b8ded277d67fe389a997a6338214d2"} Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.727574 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="728d43846cdc19398445756d1892408a23b8ded277d67fe389a997a6338214d2" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.845446 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8"] Dec 03 13:58:39 crc kubenswrapper[4690]: E1203 13:58:39.846067 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce19539-f48f-446b-8a79-d528d58a9e8d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.846090 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce19539-f48f-446b-8a79-d528d58a9e8d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.846336 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ce19539-f48f-446b-8a79-d528d58a9e8d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.847342 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.852179 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.852477 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.852638 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.852887 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.853102 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.853248 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.855218 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.860038 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8"] Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.975431 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.975834 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.976014 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.976142 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.976285 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.976399 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97p4k\" (UniqueName: \"kubernetes.io/projected/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-kube-api-access-97p4k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.976493 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.976611 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:39 crc kubenswrapper[4690]: I1203 13:58:39.976770 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.078672 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.079078 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.079113 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.079681 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97p4k\" (UniqueName: \"kubernetes.io/projected/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-kube-api-access-97p4k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.079711 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.079768 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.079957 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.080082 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.080137 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.081065 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.083392 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.083478 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.083590 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.085344 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.087164 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.089078 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.089648 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.099235 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97p4k\" (UniqueName: \"kubernetes.io/projected/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-kube-api-access-97p4k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hgcm8\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.171550 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.706404 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8"] Dec 03 13:58:40 crc kubenswrapper[4690]: I1203 13:58:40.758222 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" event={"ID":"6b06fe71-66c9-48a8-9e96-b9bdb23daf92","Type":"ContainerStarted","Data":"925ebfc53dddd03dbdcd16d9346913d2f5d613b568b758d66faa5f93c9defe72"} Dec 03 13:58:41 crc kubenswrapper[4690]: I1203 13:58:41.770419 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" event={"ID":"6b06fe71-66c9-48a8-9e96-b9bdb23daf92","Type":"ContainerStarted","Data":"aaafa31698db688df8f4b5c3e0c72c07e2edd6c9f23c55690cf1340890aabc65"} Dec 03 13:58:41 crc kubenswrapper[4690]: I1203 13:58:41.797837 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" podStartSLOduration=2.298605712 podStartE2EDuration="2.797815536s" podCreationTimestamp="2025-12-03 13:58:39 +0000 UTC" firstStartedPulling="2025-12-03 13:58:40.718497242 +0000 UTC m=+2966.699417675" lastFinishedPulling="2025-12-03 13:58:41.217707066 +0000 UTC m=+2967.198627499" observedRunningTime="2025-12-03 13:58:41.792844411 +0000 UTC m=+2967.773764844" watchObservedRunningTime="2025-12-03 13:58:41.797815536 +0000 UTC m=+2967.778735969" Dec 03 13:58:45 crc kubenswrapper[4690]: I1203 13:58:45.083137 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:45 crc kubenswrapper[4690]: I1203 13:58:45.083618 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:45 crc kubenswrapper[4690]: I1203 13:58:45.130931 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:45 crc kubenswrapper[4690]: I1203 13:58:45.871425 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:45 crc kubenswrapper[4690]: I1203 13:58:45.920203 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7slb7"] Dec 03 13:58:47 crc kubenswrapper[4690]: I1203 13:58:47.847647 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7slb7" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerName="registry-server" containerID="cri-o://02e955637cb0c6c5fd9781f4852fed62aed8c6b372c826d5e249215b898e5b41" gracePeriod=2 Dec 03 13:58:49 crc kubenswrapper[4690]: I1203 13:58:49.873272 4690 generic.go:334] "Generic (PLEG): container finished" podID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerID="02e955637cb0c6c5fd9781f4852fed62aed8c6b372c826d5e249215b898e5b41" exitCode=0 Dec 03 13:58:49 crc kubenswrapper[4690]: I1203 13:58:49.873391 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7slb7" event={"ID":"1d7d24dd-562c-4ee3-856a-57f55ba404f3","Type":"ContainerDied","Data":"02e955637cb0c6c5fd9781f4852fed62aed8c6b372c826d5e249215b898e5b41"} Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.174674 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.219333 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-catalog-content\") pod \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.219637 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shngx\" (UniqueName: \"kubernetes.io/projected/1d7d24dd-562c-4ee3-856a-57f55ba404f3-kube-api-access-shngx\") pod \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.219682 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-utilities\") pod \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\" (UID: \"1d7d24dd-562c-4ee3-856a-57f55ba404f3\") " Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.220749 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-utilities" (OuterVolumeSpecName: "utilities") pod "1d7d24dd-562c-4ee3-856a-57f55ba404f3" (UID: "1d7d24dd-562c-4ee3-856a-57f55ba404f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.231235 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7d24dd-562c-4ee3-856a-57f55ba404f3-kube-api-access-shngx" (OuterVolumeSpecName: "kube-api-access-shngx") pod "1d7d24dd-562c-4ee3-856a-57f55ba404f3" (UID: "1d7d24dd-562c-4ee3-856a-57f55ba404f3"). InnerVolumeSpecName "kube-api-access-shngx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.242753 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d7d24dd-562c-4ee3-856a-57f55ba404f3" (UID: "1d7d24dd-562c-4ee3-856a-57f55ba404f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.322311 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shngx\" (UniqueName: \"kubernetes.io/projected/1d7d24dd-562c-4ee3-856a-57f55ba404f3-kube-api-access-shngx\") on node \"crc\" DevicePath \"\"" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.322641 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.322748 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7d24dd-562c-4ee3-856a-57f55ba404f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.888081 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7slb7" event={"ID":"1d7d24dd-562c-4ee3-856a-57f55ba404f3","Type":"ContainerDied","Data":"8ea0d392a39aa9d557cb702fdeda43f56b522f8b5b807605b8b82c71dddcd25f"} Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.888146 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7slb7" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.888177 4690 scope.go:117] "RemoveContainer" containerID="02e955637cb0c6c5fd9781f4852fed62aed8c6b372c826d5e249215b898e5b41" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.926976 4690 scope.go:117] "RemoveContainer" containerID="561e985c1e2558ece5148dadcd72fa2305050daa13dd71d85c69c5deda8ed240" Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.934233 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7slb7"] Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.948401 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7slb7"] Dec 03 13:58:50 crc kubenswrapper[4690]: I1203 13:58:50.958060 4690 scope.go:117] "RemoveContainer" containerID="529df63edc24ad7931817f5233c78074c07333be77341fc965a4d8c868151b5d" Dec 03 13:58:52 crc kubenswrapper[4690]: I1203 13:58:52.328840 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" path="/var/lib/kubelet/pods/1d7d24dd-562c-4ee3-856a-57f55ba404f3/volumes" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.145379 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq"] Dec 03 14:00:00 crc kubenswrapper[4690]: E1203 14:00:00.146672 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerName="extract-utilities" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.146688 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerName="extract-utilities" Dec 03 14:00:00 crc kubenswrapper[4690]: E1203 14:00:00.146699 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerName="extract-content" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.146707 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerName="extract-content" Dec 03 14:00:00 crc kubenswrapper[4690]: E1203 14:00:00.146724 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.146730 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.146964 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d7d24dd-562c-4ee3-856a-57f55ba404f3" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.147858 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.151390 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.151450 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.160484 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq"] Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.334213 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz244\" (UniqueName: \"kubernetes.io/projected/73d269ec-26d0-4426-aa92-b0d81f20cd22-kube-api-access-zz244\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.334309 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73d269ec-26d0-4426-aa92-b0d81f20cd22-config-volume\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.334374 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73d269ec-26d0-4426-aa92-b0d81f20cd22-secret-volume\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.436347 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz244\" (UniqueName: \"kubernetes.io/projected/73d269ec-26d0-4426-aa92-b0d81f20cd22-kube-api-access-zz244\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.436447 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73d269ec-26d0-4426-aa92-b0d81f20cd22-config-volume\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.436533 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73d269ec-26d0-4426-aa92-b0d81f20cd22-secret-volume\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.437856 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73d269ec-26d0-4426-aa92-b0d81f20cd22-config-volume\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.443588 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73d269ec-26d0-4426-aa92-b0d81f20cd22-secret-volume\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.462799 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz244\" (UniqueName: \"kubernetes.io/projected/73d269ec-26d0-4426-aa92-b0d81f20cd22-kube-api-access-zz244\") pod \"collect-profiles-29412840-pnkgq\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:00 crc kubenswrapper[4690]: I1203 14:00:00.492846 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:01 crc kubenswrapper[4690]: I1203 14:00:01.020911 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq"] Dec 03 14:00:01 crc kubenswrapper[4690]: I1203 14:00:01.630979 4690 generic.go:334] "Generic (PLEG): container finished" podID="73d269ec-26d0-4426-aa92-b0d81f20cd22" containerID="0323ed75574ee6a1a9b438351cf7fae4125469f7ab48ab488b2371d72992b2d7" exitCode=0 Dec 03 14:00:01 crc kubenswrapper[4690]: I1203 14:00:01.631310 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" event={"ID":"73d269ec-26d0-4426-aa92-b0d81f20cd22","Type":"ContainerDied","Data":"0323ed75574ee6a1a9b438351cf7fae4125469f7ab48ab488b2371d72992b2d7"} Dec 03 14:00:01 crc kubenswrapper[4690]: I1203 14:00:01.631339 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" event={"ID":"73d269ec-26d0-4426-aa92-b0d81f20cd22","Type":"ContainerStarted","Data":"e4d05a1960fe8b7bd5f64b330a423be0f6907004f923920f736e682c3a84bf75"} Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.036192 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.113700 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73d269ec-26d0-4426-aa92-b0d81f20cd22-secret-volume\") pod \"73d269ec-26d0-4426-aa92-b0d81f20cd22\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.114269 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73d269ec-26d0-4426-aa92-b0d81f20cd22-config-volume\") pod \"73d269ec-26d0-4426-aa92-b0d81f20cd22\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.114405 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz244\" (UniqueName: \"kubernetes.io/projected/73d269ec-26d0-4426-aa92-b0d81f20cd22-kube-api-access-zz244\") pod \"73d269ec-26d0-4426-aa92-b0d81f20cd22\" (UID: \"73d269ec-26d0-4426-aa92-b0d81f20cd22\") " Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.115344 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73d269ec-26d0-4426-aa92-b0d81f20cd22-config-volume" (OuterVolumeSpecName: "config-volume") pod "73d269ec-26d0-4426-aa92-b0d81f20cd22" (UID: "73d269ec-26d0-4426-aa92-b0d81f20cd22"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.120544 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73d269ec-26d0-4426-aa92-b0d81f20cd22-kube-api-access-zz244" (OuterVolumeSpecName: "kube-api-access-zz244") pod "73d269ec-26d0-4426-aa92-b0d81f20cd22" (UID: "73d269ec-26d0-4426-aa92-b0d81f20cd22"). InnerVolumeSpecName "kube-api-access-zz244". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.122643 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73d269ec-26d0-4426-aa92-b0d81f20cd22-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "73d269ec-26d0-4426-aa92-b0d81f20cd22" (UID: "73d269ec-26d0-4426-aa92-b0d81f20cd22"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.215691 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73d269ec-26d0-4426-aa92-b0d81f20cd22-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.215740 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz244\" (UniqueName: \"kubernetes.io/projected/73d269ec-26d0-4426-aa92-b0d81f20cd22-kube-api-access-zz244\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.215751 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73d269ec-26d0-4426-aa92-b0d81f20cd22-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.656041 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" event={"ID":"73d269ec-26d0-4426-aa92-b0d81f20cd22","Type":"ContainerDied","Data":"e4d05a1960fe8b7bd5f64b330a423be0f6907004f923920f736e682c3a84bf75"} Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.656093 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4d05a1960fe8b7bd5f64b330a423be0f6907004f923920f736e682c3a84bf75" Dec 03 14:00:03 crc kubenswrapper[4690]: I1203 14:00:03.656097 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-pnkgq" Dec 03 14:00:04 crc kubenswrapper[4690]: I1203 14:00:04.129323 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj"] Dec 03 14:00:04 crc kubenswrapper[4690]: I1203 14:00:04.146205 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-64crj"] Dec 03 14:00:04 crc kubenswrapper[4690]: I1203 14:00:04.327767 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9235e7b-0d79-471b-96fa-08c36c7f421f" path="/var/lib/kubelet/pods/e9235e7b-0d79-471b-96fa-08c36c7f421f/volumes" Dec 03 14:00:16 crc kubenswrapper[4690]: I1203 14:00:16.824213 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:00:16 crc kubenswrapper[4690]: I1203 14:00:16.824810 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:00:42 crc kubenswrapper[4690]: I1203 14:00:42.303777 4690 scope.go:117] "RemoveContainer" containerID="20a3d77817fc21a96e4867c8a1c2452347520db2c20454b888dafbb88e6e6e01" Dec 03 14:00:46 crc kubenswrapper[4690]: I1203 14:00:46.824294 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:00:46 crc kubenswrapper[4690]: I1203 14:00:46.825032 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.153373 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29412841-vn5nx"] Dec 03 14:01:00 crc kubenswrapper[4690]: E1203 14:01:00.154590 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d269ec-26d0-4426-aa92-b0d81f20cd22" containerName="collect-profiles" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.154606 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d269ec-26d0-4426-aa92-b0d81f20cd22" containerName="collect-profiles" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.154832 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="73d269ec-26d0-4426-aa92-b0d81f20cd22" containerName="collect-profiles" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.155647 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.172394 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412841-vn5nx"] Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.252822 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-config-data\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.252903 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-combined-ca-bundle\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.252974 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-fernet-keys\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.253099 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d882s\" (UniqueName: \"kubernetes.io/projected/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-kube-api-access-d882s\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.355232 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-fernet-keys\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.355326 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d882s\" (UniqueName: \"kubernetes.io/projected/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-kube-api-access-d882s\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.355451 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-config-data\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.356582 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-combined-ca-bundle\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.363178 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-fernet-keys\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.365934 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-combined-ca-bundle\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.368518 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-config-data\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.374480 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d882s\" (UniqueName: \"kubernetes.io/projected/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-kube-api-access-d882s\") pod \"keystone-cron-29412841-vn5nx\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.483631 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:00 crc kubenswrapper[4690]: I1203 14:01:00.954936 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412841-vn5nx"] Dec 03 14:01:01 crc kubenswrapper[4690]: I1203 14:01:01.320125 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412841-vn5nx" event={"ID":"99ea4ce8-3695-4f3d-97fd-99d0e0b87029","Type":"ContainerStarted","Data":"8d7ed2a6169b978622c31ced28c42dd72175ab38219e6f60f02ec15543d0a6cc"} Dec 03 14:01:01 crc kubenswrapper[4690]: I1203 14:01:01.320550 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412841-vn5nx" event={"ID":"99ea4ce8-3695-4f3d-97fd-99d0e0b87029","Type":"ContainerStarted","Data":"0fbf71284132cea949b24b222fbd4635f6ca53e73afe616d24d6474a81330046"} Dec 03 14:01:04 crc kubenswrapper[4690]: I1203 14:01:04.356438 4690 generic.go:334] "Generic (PLEG): container finished" podID="99ea4ce8-3695-4f3d-97fd-99d0e0b87029" containerID="8d7ed2a6169b978622c31ced28c42dd72175ab38219e6f60f02ec15543d0a6cc" exitCode=0 Dec 03 14:01:04 crc kubenswrapper[4690]: I1203 14:01:04.357002 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412841-vn5nx" event={"ID":"99ea4ce8-3695-4f3d-97fd-99d0e0b87029","Type":"ContainerDied","Data":"8d7ed2a6169b978622c31ced28c42dd72175ab38219e6f60f02ec15543d0a6cc"} Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.752230 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.886184 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-config-data\") pod \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.886537 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d882s\" (UniqueName: \"kubernetes.io/projected/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-kube-api-access-d882s\") pod \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.886669 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-combined-ca-bundle\") pod \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.886694 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-fernet-keys\") pod \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\" (UID: \"99ea4ce8-3695-4f3d-97fd-99d0e0b87029\") " Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.893294 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-kube-api-access-d882s" (OuterVolumeSpecName: "kube-api-access-d882s") pod "99ea4ce8-3695-4f3d-97fd-99d0e0b87029" (UID: "99ea4ce8-3695-4f3d-97fd-99d0e0b87029"). InnerVolumeSpecName "kube-api-access-d882s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.897642 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "99ea4ce8-3695-4f3d-97fd-99d0e0b87029" (UID: "99ea4ce8-3695-4f3d-97fd-99d0e0b87029"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.917003 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99ea4ce8-3695-4f3d-97fd-99d0e0b87029" (UID: "99ea4ce8-3695-4f3d-97fd-99d0e0b87029"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.947608 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-config-data" (OuterVolumeSpecName: "config-data") pod "99ea4ce8-3695-4f3d-97fd-99d0e0b87029" (UID: "99ea4ce8-3695-4f3d-97fd-99d0e0b87029"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.989540 4690 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.989571 4690 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.989583 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:05 crc kubenswrapper[4690]: I1203 14:01:05.989594 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d882s\" (UniqueName: \"kubernetes.io/projected/99ea4ce8-3695-4f3d-97fd-99d0e0b87029-kube-api-access-d882s\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:06 crc kubenswrapper[4690]: I1203 14:01:06.378760 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412841-vn5nx" event={"ID":"99ea4ce8-3695-4f3d-97fd-99d0e0b87029","Type":"ContainerDied","Data":"0fbf71284132cea949b24b222fbd4635f6ca53e73afe616d24d6474a81330046"} Dec 03 14:01:06 crc kubenswrapper[4690]: I1203 14:01:06.378802 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fbf71284132cea949b24b222fbd4635f6ca53e73afe616d24d6474a81330046" Dec 03 14:01:06 crc kubenswrapper[4690]: I1203 14:01:06.378835 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412841-vn5nx" Dec 03 14:01:16 crc kubenswrapper[4690]: I1203 14:01:16.823516 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:01:16 crc kubenswrapper[4690]: I1203 14:01:16.824187 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:01:16 crc kubenswrapper[4690]: I1203 14:01:16.824236 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 14:01:16 crc kubenswrapper[4690]: I1203 14:01:16.825106 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:01:16 crc kubenswrapper[4690]: I1203 14:01:16.825169 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" gracePeriod=600 Dec 03 14:01:16 crc kubenswrapper[4690]: E1203 14:01:16.974589 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:01:17 crc kubenswrapper[4690]: I1203 14:01:17.488541 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" exitCode=0 Dec 03 14:01:17 crc kubenswrapper[4690]: I1203 14:01:17.488606 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615"} Dec 03 14:01:17 crc kubenswrapper[4690]: I1203 14:01:17.488662 4690 scope.go:117] "RemoveContainer" containerID="f2186a489870b5492e9f2c87bb858a3f5a82bbdd59d220ed13a4486d7275a008" Dec 03 14:01:17 crc kubenswrapper[4690]: I1203 14:01:17.489779 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:01:17 crc kubenswrapper[4690]: E1203 14:01:17.490237 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:01:29 crc kubenswrapper[4690]: I1203 14:01:29.605838 4690 generic.go:334] "Generic (PLEG): container finished" podID="6b06fe71-66c9-48a8-9e96-b9bdb23daf92" containerID="aaafa31698db688df8f4b5c3e0c72c07e2edd6c9f23c55690cf1340890aabc65" exitCode=0 Dec 03 14:01:29 crc kubenswrapper[4690]: I1203 14:01:29.605959 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" event={"ID":"6b06fe71-66c9-48a8-9e96-b9bdb23daf92","Type":"ContainerDied","Data":"aaafa31698db688df8f4b5c3e0c72c07e2edd6c9f23c55690cf1340890aabc65"} Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.096126 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.147397 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-combined-ca-bundle\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.148102 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-extra-config-0\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.148144 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-1\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.148182 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-inventory\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.148282 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97p4k\" (UniqueName: \"kubernetes.io/projected/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-kube-api-access-97p4k\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.148346 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-0\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.149095 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-0\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.149162 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-1\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.149204 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-ssh-key\") pod \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\" (UID: \"6b06fe71-66c9-48a8-9e96-b9bdb23daf92\") " Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.162214 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-kube-api-access-97p4k" (OuterVolumeSpecName: "kube-api-access-97p4k") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "kube-api-access-97p4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.173461 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.189267 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-inventory" (OuterVolumeSpecName: "inventory") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.189602 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.189705 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.191539 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.199730 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.203184 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.209165 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "6b06fe71-66c9-48a8-9e96-b9bdb23daf92" (UID: "6b06fe71-66c9-48a8-9e96-b9bdb23daf92"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253189 4690 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253237 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253247 4690 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253256 4690 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253264 4690 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253273 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253283 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97p4k\" (UniqueName: \"kubernetes.io/projected/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-kube-api-access-97p4k\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253294 4690 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.253306 4690 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b06fe71-66c9-48a8-9e96-b9bdb23daf92-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.315420 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:01:31 crc kubenswrapper[4690]: E1203 14:01:31.320615 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.630277 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" event={"ID":"6b06fe71-66c9-48a8-9e96-b9bdb23daf92","Type":"ContainerDied","Data":"925ebfc53dddd03dbdcd16d9346913d2f5d613b568b758d66faa5f93c9defe72"} Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.630655 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="925ebfc53dddd03dbdcd16d9346913d2f5d613b568b758d66faa5f93c9defe72" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.630390 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hgcm8" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.730412 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n"] Dec 03 14:01:31 crc kubenswrapper[4690]: E1203 14:01:31.731162 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ea4ce8-3695-4f3d-97fd-99d0e0b87029" containerName="keystone-cron" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.731182 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ea4ce8-3695-4f3d-97fd-99d0e0b87029" containerName="keystone-cron" Dec 03 14:01:31 crc kubenswrapper[4690]: E1203 14:01:31.731209 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b06fe71-66c9-48a8-9e96-b9bdb23daf92" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.731215 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b06fe71-66c9-48a8-9e96-b9bdb23daf92" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.731452 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ea4ce8-3695-4f3d-97fd-99d0e0b87029" containerName="keystone-cron" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.731475 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b06fe71-66c9-48a8-9e96-b9bdb23daf92" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.732254 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.734815 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.734860 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.735539 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.735596 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.737394 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.766407 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.766850 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.766927 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.766950 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.767033 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcdbh\" (UniqueName: \"kubernetes.io/projected/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-kube-api-access-qcdbh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.767077 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.767192 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.792167 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n"] Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.869260 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.869790 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.869944 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.870069 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.870255 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcdbh\" (UniqueName: \"kubernetes.io/projected/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-kube-api-access-qcdbh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.870380 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.870562 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.876550 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.876699 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.877395 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.877782 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.881505 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.893247 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:31 crc kubenswrapper[4690]: I1203 14:01:31.896701 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcdbh\" (UniqueName: \"kubernetes.io/projected/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-kube-api-access-qcdbh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:32 crc kubenswrapper[4690]: I1203 14:01:32.051623 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:01:32 crc kubenswrapper[4690]: I1203 14:01:32.568406 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n"] Dec 03 14:01:32 crc kubenswrapper[4690]: I1203 14:01:32.571600 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:01:32 crc kubenswrapper[4690]: I1203 14:01:32.642242 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" event={"ID":"c21f8ec4-b27d-4a89-bd74-e03efa0afe48","Type":"ContainerStarted","Data":"048ede4e48dcded7f04b392fad5f9e69a0b7513fb548b1fe6743ea8813812923"} Dec 03 14:01:33 crc kubenswrapper[4690]: I1203 14:01:33.654370 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" event={"ID":"c21f8ec4-b27d-4a89-bd74-e03efa0afe48","Type":"ContainerStarted","Data":"e79df99b67dedcc992fc31a2be72a4e05226e44e2a259fa0d778345f28460e1e"} Dec 03 14:01:33 crc kubenswrapper[4690]: I1203 14:01:33.682118 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" podStartSLOduration=2.294159563 podStartE2EDuration="2.682087484s" podCreationTimestamp="2025-12-03 14:01:31 +0000 UTC" firstStartedPulling="2025-12-03 14:01:32.571323713 +0000 UTC m=+3138.552244146" lastFinishedPulling="2025-12-03 14:01:32.959251614 +0000 UTC m=+3138.940172067" observedRunningTime="2025-12-03 14:01:33.670646216 +0000 UTC m=+3139.651566649" watchObservedRunningTime="2025-12-03 14:01:33.682087484 +0000 UTC m=+3139.663007917" Dec 03 14:01:45 crc kubenswrapper[4690]: I1203 14:01:45.315881 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:01:45 crc kubenswrapper[4690]: E1203 14:01:45.317005 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:01:59 crc kubenswrapper[4690]: I1203 14:01:59.315584 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:01:59 crc kubenswrapper[4690]: E1203 14:01:59.316509 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:02:13 crc kubenswrapper[4690]: I1203 14:02:13.315011 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:02:13 crc kubenswrapper[4690]: E1203 14:02:13.315979 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:02:28 crc kubenswrapper[4690]: I1203 14:02:28.315220 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:02:28 crc kubenswrapper[4690]: E1203 14:02:28.316301 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:02:40 crc kubenswrapper[4690]: I1203 14:02:40.315941 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:02:40 crc kubenswrapper[4690]: E1203 14:02:40.317045 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:02:42 crc kubenswrapper[4690]: I1203 14:02:42.408756 4690 scope.go:117] "RemoveContainer" containerID="978667c59166fabd7f88b0974895d9df0204c2289dd6cff40f72a568b1703b22" Dec 03 14:02:42 crc kubenswrapper[4690]: I1203 14:02:42.438952 4690 scope.go:117] "RemoveContainer" containerID="628b6e4d1fe5abe1dd9b643c9aeb7eb7bf28259d5d5fd4469ecf806452997088" Dec 03 14:02:42 crc kubenswrapper[4690]: I1203 14:02:42.481177 4690 scope.go:117] "RemoveContainer" containerID="6a8851103eb3243c74e3caa696f0922d6e7aa42127a7928b6ba7b460a9c22e48" Dec 03 14:02:53 crc kubenswrapper[4690]: I1203 14:02:53.315838 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:02:53 crc kubenswrapper[4690]: E1203 14:02:53.316721 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:03:07 crc kubenswrapper[4690]: I1203 14:03:07.315595 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:03:07 crc kubenswrapper[4690]: E1203 14:03:07.316988 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:03:18 crc kubenswrapper[4690]: I1203 14:03:18.315186 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:03:18 crc kubenswrapper[4690]: E1203 14:03:18.316050 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:03:29 crc kubenswrapper[4690]: I1203 14:03:29.315396 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:03:29 crc kubenswrapper[4690]: E1203 14:03:29.316329 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:03:43 crc kubenswrapper[4690]: I1203 14:03:43.315907 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:03:43 crc kubenswrapper[4690]: E1203 14:03:43.316984 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:03:54 crc kubenswrapper[4690]: I1203 14:03:54.321440 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:03:54 crc kubenswrapper[4690]: E1203 14:03:54.322635 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:04:03 crc kubenswrapper[4690]: I1203 14:04:03.212583 4690 generic.go:334] "Generic (PLEG): container finished" podID="c21f8ec4-b27d-4a89-bd74-e03efa0afe48" containerID="e79df99b67dedcc992fc31a2be72a4e05226e44e2a259fa0d778345f28460e1e" exitCode=0 Dec 03 14:04:03 crc kubenswrapper[4690]: I1203 14:04:03.212678 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" event={"ID":"c21f8ec4-b27d-4a89-bd74-e03efa0afe48","Type":"ContainerDied","Data":"e79df99b67dedcc992fc31a2be72a4e05226e44e2a259fa0d778345f28460e1e"} Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.713333 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.896177 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-2\") pod \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.896525 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-inventory\") pod \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.896579 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-0\") pod \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.896600 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcdbh\" (UniqueName: \"kubernetes.io/projected/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-kube-api-access-qcdbh\") pod \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.896635 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ssh-key\") pod \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.896755 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-telemetry-combined-ca-bundle\") pod \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.896795 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-1\") pod \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\" (UID: \"c21f8ec4-b27d-4a89-bd74-e03efa0afe48\") " Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.902100 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-kube-api-access-qcdbh" (OuterVolumeSpecName: "kube-api-access-qcdbh") pod "c21f8ec4-b27d-4a89-bd74-e03efa0afe48" (UID: "c21f8ec4-b27d-4a89-bd74-e03efa0afe48"). InnerVolumeSpecName "kube-api-access-qcdbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.903033 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c21f8ec4-b27d-4a89-bd74-e03efa0afe48" (UID: "c21f8ec4-b27d-4a89-bd74-e03efa0afe48"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.929646 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c21f8ec4-b27d-4a89-bd74-e03efa0afe48" (UID: "c21f8ec4-b27d-4a89-bd74-e03efa0afe48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.930325 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "c21f8ec4-b27d-4a89-bd74-e03efa0afe48" (UID: "c21f8ec4-b27d-4a89-bd74-e03efa0afe48"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.940442 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-inventory" (OuterVolumeSpecName: "inventory") pod "c21f8ec4-b27d-4a89-bd74-e03efa0afe48" (UID: "c21f8ec4-b27d-4a89-bd74-e03efa0afe48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.940900 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "c21f8ec4-b27d-4a89-bd74-e03efa0afe48" (UID: "c21f8ec4-b27d-4a89-bd74-e03efa0afe48"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:04 crc kubenswrapper[4690]: I1203 14:04:04.944718 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "c21f8ec4-b27d-4a89-bd74-e03efa0afe48" (UID: "c21f8ec4-b27d-4a89-bd74-e03efa0afe48"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.000296 4690 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.000348 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.000364 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcdbh\" (UniqueName: \"kubernetes.io/projected/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-kube-api-access-qcdbh\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.000384 4690 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.000398 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.000412 4690 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.000424 4690 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c21f8ec4-b27d-4a89-bd74-e03efa0afe48-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.235543 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" event={"ID":"c21f8ec4-b27d-4a89-bd74-e03efa0afe48","Type":"ContainerDied","Data":"048ede4e48dcded7f04b392fad5f9e69a0b7513fb548b1fe6743ea8813812923"} Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.235593 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="048ede4e48dcded7f04b392fad5f9e69a0b7513fb548b1fe6743ea8813812923" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.235653 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.358420 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl"] Dec 03 14:04:05 crc kubenswrapper[4690]: E1203 14:04:05.359039 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c21f8ec4-b27d-4a89-bd74-e03efa0afe48" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.359065 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c21f8ec4-b27d-4a89-bd74-e03efa0afe48" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.359339 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c21f8ec4-b27d-4a89-bd74-e03efa0afe48" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.360356 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.362556 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.362951 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.364644 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.364716 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.365035 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-85ztj" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.371457 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl"] Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.511173 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s28s9\" (UniqueName: \"kubernetes.io/projected/f66f8700-77ac-4bed-b21b-8866f0edc58b-kube-api-access-s28s9\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.511246 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.511332 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.511436 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.511494 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.613477 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s28s9\" (UniqueName: \"kubernetes.io/projected/f66f8700-77ac-4bed-b21b-8866f0edc58b-kube-api-access-s28s9\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.613550 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.613591 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.613663 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.613688 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.617760 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.617799 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.617826 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.625849 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.634412 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s28s9\" (UniqueName: \"kubernetes.io/projected/f66f8700-77ac-4bed-b21b-8866f0edc58b-kube-api-access-s28s9\") pod \"logging-edpm-deployment-openstack-edpm-ipam-tgvwl\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:05 crc kubenswrapper[4690]: I1203 14:04:05.680013 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:06 crc kubenswrapper[4690]: I1203 14:04:06.233915 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl"] Dec 03 14:04:06 crc kubenswrapper[4690]: I1203 14:04:06.256367 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" event={"ID":"f66f8700-77ac-4bed-b21b-8866f0edc58b","Type":"ContainerStarted","Data":"b326e8c3899079380b7926f2f0f3b407de6507f87c58cfdb228bec2b3bdc8e21"} Dec 03 14:04:06 crc kubenswrapper[4690]: I1203 14:04:06.315829 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:04:06 crc kubenswrapper[4690]: E1203 14:04:06.316220 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:04:07 crc kubenswrapper[4690]: I1203 14:04:07.271617 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" event={"ID":"f66f8700-77ac-4bed-b21b-8866f0edc58b","Type":"ContainerStarted","Data":"cc374786700b4245b6c56857cb8b7302b3b3c8cae04ca5b00b3718b2b7c1ebf2"} Dec 03 14:04:07 crc kubenswrapper[4690]: I1203 14:04:07.299682 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" podStartSLOduration=1.764171683 podStartE2EDuration="2.299659449s" podCreationTimestamp="2025-12-03 14:04:05 +0000 UTC" firstStartedPulling="2025-12-03 14:04:06.247464169 +0000 UTC m=+3292.228384602" lastFinishedPulling="2025-12-03 14:04:06.782951935 +0000 UTC m=+3292.763872368" observedRunningTime="2025-12-03 14:04:07.29492149 +0000 UTC m=+3293.275841923" watchObservedRunningTime="2025-12-03 14:04:07.299659449 +0000 UTC m=+3293.280579882" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.015392 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lgw8x"] Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.018248 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.035953 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgw8x"] Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.094345 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc29w\" (UniqueName: \"kubernetes.io/projected/bd39d84c-a392-407c-872b-9be544eb1816-kube-api-access-rc29w\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.094529 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd39d84c-a392-407c-872b-9be544eb1816-catalog-content\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.094641 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd39d84c-a392-407c-872b-9be544eb1816-utilities\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.197391 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd39d84c-a392-407c-872b-9be544eb1816-catalog-content\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.197495 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd39d84c-a392-407c-872b-9be544eb1816-utilities\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.197612 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc29w\" (UniqueName: \"kubernetes.io/projected/bd39d84c-a392-407c-872b-9be544eb1816-kube-api-access-rc29w\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.198151 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd39d84c-a392-407c-872b-9be544eb1816-catalog-content\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.198252 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd39d84c-a392-407c-872b-9be544eb1816-utilities\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.227764 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc29w\" (UniqueName: \"kubernetes.io/projected/bd39d84c-a392-407c-872b-9be544eb1816-kube-api-access-rc29w\") pod \"redhat-operators-lgw8x\" (UID: \"bd39d84c-a392-407c-872b-9be544eb1816\") " pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.344446 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:12 crc kubenswrapper[4690]: I1203 14:04:12.893410 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgw8x"] Dec 03 14:04:13 crc kubenswrapper[4690]: I1203 14:04:13.332652 4690 generic.go:334] "Generic (PLEG): container finished" podID="bd39d84c-a392-407c-872b-9be544eb1816" containerID="869915add1e86c96509b225210cad7a337c4119ab95489c6a2872bb5da1830a6" exitCode=0 Dec 03 14:04:13 crc kubenswrapper[4690]: I1203 14:04:13.332706 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgw8x" event={"ID":"bd39d84c-a392-407c-872b-9be544eb1816","Type":"ContainerDied","Data":"869915add1e86c96509b225210cad7a337c4119ab95489c6a2872bb5da1830a6"} Dec 03 14:04:13 crc kubenswrapper[4690]: I1203 14:04:13.332738 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgw8x" event={"ID":"bd39d84c-a392-407c-872b-9be544eb1816","Type":"ContainerStarted","Data":"e6083fed34b574826f87ff7c3532ec4fbb9fc33f1be752211a4fe7b6717a1c5a"} Dec 03 14:04:21 crc kubenswrapper[4690]: I1203 14:04:21.314989 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:04:21 crc kubenswrapper[4690]: E1203 14:04:21.315997 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:04:21 crc kubenswrapper[4690]: I1203 14:04:21.448400 4690 generic.go:334] "Generic (PLEG): container finished" podID="f66f8700-77ac-4bed-b21b-8866f0edc58b" containerID="cc374786700b4245b6c56857cb8b7302b3b3c8cae04ca5b00b3718b2b7c1ebf2" exitCode=0 Dec 03 14:04:21 crc kubenswrapper[4690]: I1203 14:04:21.448454 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" event={"ID":"f66f8700-77ac-4bed-b21b-8866f0edc58b","Type":"ContainerDied","Data":"cc374786700b4245b6c56857cb8b7302b3b3c8cae04ca5b00b3718b2b7c1ebf2"} Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.021594 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.176153 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-ssh-key\") pod \"f66f8700-77ac-4bed-b21b-8866f0edc58b\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.176281 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-1\") pod \"f66f8700-77ac-4bed-b21b-8866f0edc58b\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.176512 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-inventory\") pod \"f66f8700-77ac-4bed-b21b-8866f0edc58b\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.176596 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s28s9\" (UniqueName: \"kubernetes.io/projected/f66f8700-77ac-4bed-b21b-8866f0edc58b-kube-api-access-s28s9\") pod \"f66f8700-77ac-4bed-b21b-8866f0edc58b\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.176713 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-0\") pod \"f66f8700-77ac-4bed-b21b-8866f0edc58b\" (UID: \"f66f8700-77ac-4bed-b21b-8866f0edc58b\") " Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.192267 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f66f8700-77ac-4bed-b21b-8866f0edc58b-kube-api-access-s28s9" (OuterVolumeSpecName: "kube-api-access-s28s9") pod "f66f8700-77ac-4bed-b21b-8866f0edc58b" (UID: "f66f8700-77ac-4bed-b21b-8866f0edc58b"). InnerVolumeSpecName "kube-api-access-s28s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.207966 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "f66f8700-77ac-4bed-b21b-8866f0edc58b" (UID: "f66f8700-77ac-4bed-b21b-8866f0edc58b"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.208421 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-inventory" (OuterVolumeSpecName: "inventory") pod "f66f8700-77ac-4bed-b21b-8866f0edc58b" (UID: "f66f8700-77ac-4bed-b21b-8866f0edc58b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.210488 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "f66f8700-77ac-4bed-b21b-8866f0edc58b" (UID: "f66f8700-77ac-4bed-b21b-8866f0edc58b"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.213935 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f66f8700-77ac-4bed-b21b-8866f0edc58b" (UID: "f66f8700-77ac-4bed-b21b-8866f0edc58b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.279932 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s28s9\" (UniqueName: \"kubernetes.io/projected/f66f8700-77ac-4bed-b21b-8866f0edc58b-kube-api-access-s28s9\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.279980 4690 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.279994 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.280006 4690 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.280033 4690 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f66f8700-77ac-4bed-b21b-8866f0edc58b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.470467 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" event={"ID":"f66f8700-77ac-4bed-b21b-8866f0edc58b","Type":"ContainerDied","Data":"b326e8c3899079380b7926f2f0f3b407de6507f87c58cfdb228bec2b3bdc8e21"} Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.470512 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b326e8c3899079380b7926f2f0f3b407de6507f87c58cfdb228bec2b3bdc8e21" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.470575 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-tgvwl" Dec 03 14:04:23 crc kubenswrapper[4690]: I1203 14:04:23.473345 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgw8x" event={"ID":"bd39d84c-a392-407c-872b-9be544eb1816","Type":"ContainerStarted","Data":"7d73b7e4651f1fcb533f3644f1b63f7c43d2f08a738426a458d82d5f518867a5"} Dec 03 14:04:25 crc kubenswrapper[4690]: I1203 14:04:25.497466 4690 generic.go:334] "Generic (PLEG): container finished" podID="bd39d84c-a392-407c-872b-9be544eb1816" containerID="7d73b7e4651f1fcb533f3644f1b63f7c43d2f08a738426a458d82d5f518867a5" exitCode=0 Dec 03 14:04:25 crc kubenswrapper[4690]: I1203 14:04:25.497566 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgw8x" event={"ID":"bd39d84c-a392-407c-872b-9be544eb1816","Type":"ContainerDied","Data":"7d73b7e4651f1fcb533f3644f1b63f7c43d2f08a738426a458d82d5f518867a5"} Dec 03 14:04:31 crc kubenswrapper[4690]: I1203 14:04:31.554941 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgw8x" event={"ID":"bd39d84c-a392-407c-872b-9be544eb1816","Type":"ContainerStarted","Data":"ae86d4a2696c5affcd0381d1dcc91d31d307c9652233e55b726eb01e738df900"} Dec 03 14:04:32 crc kubenswrapper[4690]: I1203 14:04:32.314712 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:04:32 crc kubenswrapper[4690]: E1203 14:04:32.315309 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:04:32 crc kubenswrapper[4690]: I1203 14:04:32.591794 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lgw8x" podStartSLOduration=4.526720415 podStartE2EDuration="21.5917651s" podCreationTimestamp="2025-12-03 14:04:11 +0000 UTC" firstStartedPulling="2025-12-03 14:04:13.334456644 +0000 UTC m=+3299.315377077" lastFinishedPulling="2025-12-03 14:04:30.399501329 +0000 UTC m=+3316.380421762" observedRunningTime="2025-12-03 14:04:32.583208575 +0000 UTC m=+3318.564128998" watchObservedRunningTime="2025-12-03 14:04:32.5917651 +0000 UTC m=+3318.572685533" Dec 03 14:04:42 crc kubenswrapper[4690]: I1203 14:04:42.345509 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:42 crc kubenswrapper[4690]: I1203 14:04:42.346455 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:42 crc kubenswrapper[4690]: I1203 14:04:42.404458 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:42 crc kubenswrapper[4690]: I1203 14:04:42.729046 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lgw8x" Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.042428 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgw8x"] Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.223199 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p4jzs"] Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.223810 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p4jzs" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="registry-server" containerID="cri-o://81d4d3c516acfacedd8298b5b036aa71df6234922824d3a4819d535cbbbad77c" gracePeriod=2 Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.695073 4690 generic.go:334] "Generic (PLEG): container finished" podID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerID="81d4d3c516acfacedd8298b5b036aa71df6234922824d3a4819d535cbbbad77c" exitCode=0 Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.696209 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4jzs" event={"ID":"3d01a069-e06b-42e6-9667-0c8e57306a2a","Type":"ContainerDied","Data":"81d4d3c516acfacedd8298b5b036aa71df6234922824d3a4819d535cbbbad77c"} Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.820353 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.927887 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-catalog-content\") pod \"3d01a069-e06b-42e6-9667-0c8e57306a2a\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.927965 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-utilities\") pod \"3d01a069-e06b-42e6-9667-0c8e57306a2a\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.928127 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq7nh\" (UniqueName: \"kubernetes.io/projected/3d01a069-e06b-42e6-9667-0c8e57306a2a-kube-api-access-rq7nh\") pod \"3d01a069-e06b-42e6-9667-0c8e57306a2a\" (UID: \"3d01a069-e06b-42e6-9667-0c8e57306a2a\") " Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.929749 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-utilities" (OuterVolumeSpecName: "utilities") pod "3d01a069-e06b-42e6-9667-0c8e57306a2a" (UID: "3d01a069-e06b-42e6-9667-0c8e57306a2a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:04:43 crc kubenswrapper[4690]: I1203 14:04:43.959115 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d01a069-e06b-42e6-9667-0c8e57306a2a-kube-api-access-rq7nh" (OuterVolumeSpecName: "kube-api-access-rq7nh") pod "3d01a069-e06b-42e6-9667-0c8e57306a2a" (UID: "3d01a069-e06b-42e6-9667-0c8e57306a2a"). InnerVolumeSpecName "kube-api-access-rq7nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.032498 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq7nh\" (UniqueName: \"kubernetes.io/projected/3d01a069-e06b-42e6-9667-0c8e57306a2a-kube-api-access-rq7nh\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.032553 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.213491 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d01a069-e06b-42e6-9667-0c8e57306a2a" (UID: "3d01a069-e06b-42e6-9667-0c8e57306a2a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.241728 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d01a069-e06b-42e6-9667-0c8e57306a2a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.709106 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4jzs" event={"ID":"3d01a069-e06b-42e6-9667-0c8e57306a2a","Type":"ContainerDied","Data":"0915cf7e938e0fb8a9a55ad0b0f0e35e34befcb3c9c81bc7302e04d488438799"} Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.709263 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4jzs" Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.709587 4690 scope.go:117] "RemoveContainer" containerID="81d4d3c516acfacedd8298b5b036aa71df6234922824d3a4819d535cbbbad77c" Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.741197 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p4jzs"] Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.752234 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p4jzs"] Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.759268 4690 scope.go:117] "RemoveContainer" containerID="ae5c1135e3570e449535946c2f1abb29081239f608c8b454c1a2efa92ca907a5" Dec 03 14:04:44 crc kubenswrapper[4690]: I1203 14:04:44.791088 4690 scope.go:117] "RemoveContainer" containerID="33ec16a8131c112a06881b898fd70a6d0821e42c45b976cc7a7684c605ccfb58" Dec 03 14:04:45 crc kubenswrapper[4690]: I1203 14:04:45.314803 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:04:45 crc kubenswrapper[4690]: E1203 14:04:45.315210 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:04:46 crc kubenswrapper[4690]: I1203 14:04:46.327497 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" path="/var/lib/kubelet/pods/3d01a069-e06b-42e6-9667-0c8e57306a2a/volumes" Dec 03 14:04:56 crc kubenswrapper[4690]: I1203 14:04:56.317647 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:04:56 crc kubenswrapper[4690]: E1203 14:04:56.325440 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:05:09 crc kubenswrapper[4690]: I1203 14:05:09.315372 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:05:09 crc kubenswrapper[4690]: E1203 14:05:09.316370 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:05:24 crc kubenswrapper[4690]: I1203 14:05:24.321033 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:05:24 crc kubenswrapper[4690]: E1203 14:05:24.321935 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:05:37 crc kubenswrapper[4690]: I1203 14:05:37.314604 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:05:37 crc kubenswrapper[4690]: E1203 14:05:37.315575 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:05:52 crc kubenswrapper[4690]: I1203 14:05:52.315635 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:05:52 crc kubenswrapper[4690]: E1203 14:05:52.316578 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:06:06 crc kubenswrapper[4690]: I1203 14:06:06.315592 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:06:06 crc kubenswrapper[4690]: E1203 14:06:06.316471 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.882987 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rn9lq"] Dec 03 14:06:16 crc kubenswrapper[4690]: E1203 14:06:16.884237 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66f8700-77ac-4bed-b21b-8866f0edc58b" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.884256 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66f8700-77ac-4bed-b21b-8866f0edc58b" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 14:06:16 crc kubenswrapper[4690]: E1203 14:06:16.884272 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="extract-utilities" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.884278 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="extract-utilities" Dec 03 14:06:16 crc kubenswrapper[4690]: E1203 14:06:16.884302 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="extract-content" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.884308 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="extract-content" Dec 03 14:06:16 crc kubenswrapper[4690]: E1203 14:06:16.884339 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="registry-server" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.884346 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="registry-server" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.884608 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d01a069-e06b-42e6-9667-0c8e57306a2a" containerName="registry-server" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.884625 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66f8700-77ac-4bed-b21b-8866f0edc58b" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.886352 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.907530 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddkpj\" (UniqueName: \"kubernetes.io/projected/c24731a2-cf17-47b7-bded-4e2c01c63ce3-kube-api-access-ddkpj\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.907701 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-utilities\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.907745 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-catalog-content\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:16 crc kubenswrapper[4690]: I1203 14:06:16.913955 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rn9lq"] Dec 03 14:06:17 crc kubenswrapper[4690]: I1203 14:06:17.009324 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-catalog-content\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:17 crc kubenswrapper[4690]: I1203 14:06:17.009725 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddkpj\" (UniqueName: \"kubernetes.io/projected/c24731a2-cf17-47b7-bded-4e2c01c63ce3-kube-api-access-ddkpj\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:17 crc kubenswrapper[4690]: I1203 14:06:17.009922 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-utilities\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:17 crc kubenswrapper[4690]: I1203 14:06:17.010105 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-catalog-content\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:17 crc kubenswrapper[4690]: I1203 14:06:17.010416 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-utilities\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:17 crc kubenswrapper[4690]: I1203 14:06:17.031420 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddkpj\" (UniqueName: \"kubernetes.io/projected/c24731a2-cf17-47b7-bded-4e2c01c63ce3-kube-api-access-ddkpj\") pod \"community-operators-rn9lq\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:17 crc kubenswrapper[4690]: I1203 14:06:17.220733 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:17 crc kubenswrapper[4690]: I1203 14:06:17.770736 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rn9lq"] Dec 03 14:06:18 crc kubenswrapper[4690]: I1203 14:06:18.688205 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn9lq" event={"ID":"c24731a2-cf17-47b7-bded-4e2c01c63ce3","Type":"ContainerStarted","Data":"ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5"} Dec 03 14:06:18 crc kubenswrapper[4690]: I1203 14:06:18.688590 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn9lq" event={"ID":"c24731a2-cf17-47b7-bded-4e2c01c63ce3","Type":"ContainerStarted","Data":"c4034e14dcb8ae05f372d249d0a863d15962ef2ebd1c136645460d511d3cce78"} Dec 03 14:06:19 crc kubenswrapper[4690]: I1203 14:06:19.700837 4690 generic.go:334] "Generic (PLEG): container finished" podID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerID="ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5" exitCode=0 Dec 03 14:06:19 crc kubenswrapper[4690]: I1203 14:06:19.701065 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn9lq" event={"ID":"c24731a2-cf17-47b7-bded-4e2c01c63ce3","Type":"ContainerDied","Data":"ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5"} Dec 03 14:06:20 crc kubenswrapper[4690]: I1203 14:06:20.713348 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn9lq" event={"ID":"c24731a2-cf17-47b7-bded-4e2c01c63ce3","Type":"ContainerStarted","Data":"dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c"} Dec 03 14:06:21 crc kubenswrapper[4690]: I1203 14:06:21.314931 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:06:21 crc kubenswrapper[4690]: I1203 14:06:21.726294 4690 generic.go:334] "Generic (PLEG): container finished" podID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerID="dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c" exitCode=0 Dec 03 14:06:21 crc kubenswrapper[4690]: I1203 14:06:21.726516 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn9lq" event={"ID":"c24731a2-cf17-47b7-bded-4e2c01c63ce3","Type":"ContainerDied","Data":"dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c"} Dec 03 14:06:21 crc kubenswrapper[4690]: I1203 14:06:21.730742 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"0c88144a8ab40e991dd2e75e0df62f1bf2a23d73e39a209a9eec846dd271c877"} Dec 03 14:06:22 crc kubenswrapper[4690]: I1203 14:06:22.744044 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn9lq" event={"ID":"c24731a2-cf17-47b7-bded-4e2c01c63ce3","Type":"ContainerStarted","Data":"4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0"} Dec 03 14:06:22 crc kubenswrapper[4690]: I1203 14:06:22.771532 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rn9lq" podStartSLOduration=4.308723575 podStartE2EDuration="6.771508369s" podCreationTimestamp="2025-12-03 14:06:16 +0000 UTC" firstStartedPulling="2025-12-03 14:06:19.704462123 +0000 UTC m=+3425.685382566" lastFinishedPulling="2025-12-03 14:06:22.167246927 +0000 UTC m=+3428.148167360" observedRunningTime="2025-12-03 14:06:22.766993726 +0000 UTC m=+3428.747914159" watchObservedRunningTime="2025-12-03 14:06:22.771508369 +0000 UTC m=+3428.752428802" Dec 03 14:06:27 crc kubenswrapper[4690]: I1203 14:06:27.221168 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:27 crc kubenswrapper[4690]: I1203 14:06:27.221770 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:27 crc kubenswrapper[4690]: I1203 14:06:27.280645 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:27 crc kubenswrapper[4690]: I1203 14:06:27.858861 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:27 crc kubenswrapper[4690]: I1203 14:06:27.918101 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rn9lq"] Dec 03 14:06:29 crc kubenswrapper[4690]: I1203 14:06:29.830659 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rn9lq" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerName="registry-server" containerID="cri-o://4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0" gracePeriod=2 Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.432862 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.536482 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddkpj\" (UniqueName: \"kubernetes.io/projected/c24731a2-cf17-47b7-bded-4e2c01c63ce3-kube-api-access-ddkpj\") pod \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.536672 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-catalog-content\") pod \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.536793 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-utilities\") pod \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\" (UID: \"c24731a2-cf17-47b7-bded-4e2c01c63ce3\") " Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.538153 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-utilities" (OuterVolumeSpecName: "utilities") pod "c24731a2-cf17-47b7-bded-4e2c01c63ce3" (UID: "c24731a2-cf17-47b7-bded-4e2c01c63ce3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.544754 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c24731a2-cf17-47b7-bded-4e2c01c63ce3-kube-api-access-ddkpj" (OuterVolumeSpecName: "kube-api-access-ddkpj") pod "c24731a2-cf17-47b7-bded-4e2c01c63ce3" (UID: "c24731a2-cf17-47b7-bded-4e2c01c63ce3"). InnerVolumeSpecName "kube-api-access-ddkpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.596762 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c24731a2-cf17-47b7-bded-4e2c01c63ce3" (UID: "c24731a2-cf17-47b7-bded-4e2c01c63ce3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.642283 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddkpj\" (UniqueName: \"kubernetes.io/projected/c24731a2-cf17-47b7-bded-4e2c01c63ce3-kube-api-access-ddkpj\") on node \"crc\" DevicePath \"\"" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.642328 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.642336 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c24731a2-cf17-47b7-bded-4e2c01c63ce3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.847504 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn9lq" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.848404 4690 generic.go:334] "Generic (PLEG): container finished" podID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerID="4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0" exitCode=0 Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.847510 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn9lq" event={"ID":"c24731a2-cf17-47b7-bded-4e2c01c63ce3","Type":"ContainerDied","Data":"4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0"} Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.849034 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn9lq" event={"ID":"c24731a2-cf17-47b7-bded-4e2c01c63ce3","Type":"ContainerDied","Data":"c4034e14dcb8ae05f372d249d0a863d15962ef2ebd1c136645460d511d3cce78"} Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.849159 4690 scope.go:117] "RemoveContainer" containerID="4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.874233 4690 scope.go:117] "RemoveContainer" containerID="dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.889161 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rn9lq"] Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.903850 4690 scope.go:117] "RemoveContainer" containerID="ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.905418 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rn9lq"] Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.956644 4690 scope.go:117] "RemoveContainer" containerID="4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0" Dec 03 14:06:30 crc kubenswrapper[4690]: E1203 14:06:30.957257 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0\": container with ID starting with 4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0 not found: ID does not exist" containerID="4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.957302 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0"} err="failed to get container status \"4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0\": rpc error: code = NotFound desc = could not find container \"4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0\": container with ID starting with 4d5be2a60fbdc15dd0a59097819459023c7eaa08d7c90102f13dd00d6b97a2a0 not found: ID does not exist" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.957331 4690 scope.go:117] "RemoveContainer" containerID="dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c" Dec 03 14:06:30 crc kubenswrapper[4690]: E1203 14:06:30.957635 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c\": container with ID starting with dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c not found: ID does not exist" containerID="dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.957697 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c"} err="failed to get container status \"dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c\": rpc error: code = NotFound desc = could not find container \"dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c\": container with ID starting with dc7c115db11e9621b10efafae6098ef353244d1d4108f4eb873070c2a282002c not found: ID does not exist" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.957724 4690 scope.go:117] "RemoveContainer" containerID="ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5" Dec 03 14:06:30 crc kubenswrapper[4690]: E1203 14:06:30.958155 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5\": container with ID starting with ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5 not found: ID does not exist" containerID="ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5" Dec 03 14:06:30 crc kubenswrapper[4690]: I1203 14:06:30.958178 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5"} err="failed to get container status \"ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5\": rpc error: code = NotFound desc = could not find container \"ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5\": container with ID starting with ba14ec0296beb212e0e4da5a85d168cf79c766654bf8390a50d8fb4f9b716ba5 not found: ID does not exist" Dec 03 14:06:32 crc kubenswrapper[4690]: I1203 14:06:32.330061 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" path="/var/lib/kubelet/pods/c24731a2-cf17-47b7-bded-4e2c01c63ce3/volumes" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.752603 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 14:07:25 crc kubenswrapper[4690]: E1203 14:07:25.753812 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerName="extract-content" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.753827 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerName="extract-content" Dec 03 14:07:25 crc kubenswrapper[4690]: E1203 14:07:25.753842 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerName="registry-server" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.753848 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerName="registry-server" Dec 03 14:07:25 crc kubenswrapper[4690]: E1203 14:07:25.753894 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerName="extract-utilities" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.753900 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerName="extract-utilities" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.754097 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24731a2-cf17-47b7-bded-4e2c01c63ce3" containerName="registry-server" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.754934 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.763653 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.769905 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cbk7k" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.770215 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.770426 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.772068 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.819373 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.821368 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8f6w\" (UniqueName: \"kubernetes.io/projected/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-kube-api-access-r8f6w\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.821444 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.821698 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.821760 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.821920 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.822042 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.822079 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.822139 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-config-data\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.924650 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.924765 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.924802 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.924839 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.924897 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.925019 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.925075 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-config-data\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.925232 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.925379 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8f6w\" (UniqueName: \"kubernetes.io/projected/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-kube-api-access-r8f6w\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.927976 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-config-data\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.933224 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.933468 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.933525 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.936673 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.941531 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.942492 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.964196 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8f6w\" (UniqueName: \"kubernetes.io/projected/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-kube-api-access-r8f6w\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.964669 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:25 crc kubenswrapper[4690]: I1203 14:07:25.973188 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " pod="openstack/tempest-tests-tempest" Dec 03 14:07:26 crc kubenswrapper[4690]: I1203 14:07:26.102625 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 14:07:26 crc kubenswrapper[4690]: I1203 14:07:26.559435 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 14:07:26 crc kubenswrapper[4690]: I1203 14:07:26.577994 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:07:27 crc kubenswrapper[4690]: I1203 14:07:27.396280 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5","Type":"ContainerStarted","Data":"859b71dcf00cbc746d37fc58876e5a0339190023821a9c978bb6b424f9d2d944"} Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.738059 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-28nm9"] Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.742455 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.778510 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-28nm9"] Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.813928 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5tpd\" (UniqueName: \"kubernetes.io/projected/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-kube-api-access-t5tpd\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.814019 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-catalog-content\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.814119 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-utilities\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.916050 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-catalog-content\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.916205 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-utilities\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.916308 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5tpd\" (UniqueName: \"kubernetes.io/projected/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-kube-api-access-t5tpd\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.916789 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-catalog-content\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.916823 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-utilities\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:37 crc kubenswrapper[4690]: I1203 14:07:37.947133 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5tpd\" (UniqueName: \"kubernetes.io/projected/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-kube-api-access-t5tpd\") pod \"certified-operators-28nm9\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:38 crc kubenswrapper[4690]: I1203 14:07:38.077663 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:07:54 crc kubenswrapper[4690]: E1203 14:07:54.594707 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 03 14:07:54 crc kubenswrapper[4690]: E1203 14:07:54.595489 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r8f6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(46b67a35-a83b-48dd-a6ea-4464ff4ee3d5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:07:54 crc kubenswrapper[4690]: E1203 14:07:54.596756 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" Dec 03 14:07:54 crc kubenswrapper[4690]: E1203 14:07:54.726191 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" Dec 03 14:07:54 crc kubenswrapper[4690]: I1203 14:07:54.796997 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-28nm9"] Dec 03 14:07:55 crc kubenswrapper[4690]: I1203 14:07:55.736489 4690 generic.go:334] "Generic (PLEG): container finished" podID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerID="3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8" exitCode=0 Dec 03 14:07:55 crc kubenswrapper[4690]: I1203 14:07:55.736598 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28nm9" event={"ID":"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f","Type":"ContainerDied","Data":"3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8"} Dec 03 14:07:55 crc kubenswrapper[4690]: I1203 14:07:55.737940 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28nm9" event={"ID":"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f","Type":"ContainerStarted","Data":"cc5294adc15597559ecd3d45b9b02a5326d890a2731cab9a10f2e211fa4087c5"} Dec 03 14:07:57 crc kubenswrapper[4690]: E1203 14:07:57.373176 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65f7c58a_f5a3_4c47_a7c2_c50f9de3395f.slice/crio-conmon-893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:07:57 crc kubenswrapper[4690]: I1203 14:07:57.759648 4690 generic.go:334] "Generic (PLEG): container finished" podID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerID="893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49" exitCode=0 Dec 03 14:07:57 crc kubenswrapper[4690]: I1203 14:07:57.759725 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28nm9" event={"ID":"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f","Type":"ContainerDied","Data":"893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49"} Dec 03 14:07:59 crc kubenswrapper[4690]: I1203 14:07:59.784162 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28nm9" event={"ID":"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f","Type":"ContainerStarted","Data":"e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b"} Dec 03 14:07:59 crc kubenswrapper[4690]: I1203 14:07:59.833128 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-28nm9" podStartSLOduration=19.271757877 podStartE2EDuration="22.833097322s" podCreationTimestamp="2025-12-03 14:07:37 +0000 UTC" firstStartedPulling="2025-12-03 14:07:55.739458873 +0000 UTC m=+3521.720379306" lastFinishedPulling="2025-12-03 14:07:59.300798318 +0000 UTC m=+3525.281718751" observedRunningTime="2025-12-03 14:07:59.812923846 +0000 UTC m=+3525.793844279" watchObservedRunningTime="2025-12-03 14:07:59.833097322 +0000 UTC m=+3525.814017765" Dec 03 14:08:08 crc kubenswrapper[4690]: I1203 14:08:08.078413 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:08:08 crc kubenswrapper[4690]: I1203 14:08:08.080252 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:08:08 crc kubenswrapper[4690]: I1203 14:08:08.128670 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:08:08 crc kubenswrapper[4690]: I1203 14:08:08.964955 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:08:09 crc kubenswrapper[4690]: I1203 14:08:09.021024 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-28nm9"] Dec 03 14:08:10 crc kubenswrapper[4690]: I1203 14:08:10.956960 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5","Type":"ContainerStarted","Data":"d01735d62d236c5097c74d0a792ce3463f51495439f7e412e1dd93f87ebfb4d2"} Dec 03 14:08:10 crc kubenswrapper[4690]: I1203 14:08:10.957218 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-28nm9" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerName="registry-server" containerID="cri-o://e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b" gracePeriod=2 Dec 03 14:08:10 crc kubenswrapper[4690]: I1203 14:08:10.993279 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.81021164 podStartE2EDuration="46.993255102s" podCreationTimestamp="2025-12-03 14:07:24 +0000 UTC" firstStartedPulling="2025-12-03 14:07:26.577734156 +0000 UTC m=+3492.558654589" lastFinishedPulling="2025-12-03 14:08:09.760777618 +0000 UTC m=+3535.741698051" observedRunningTime="2025-12-03 14:08:10.991945409 +0000 UTC m=+3536.972865862" watchObservedRunningTime="2025-12-03 14:08:10.993255102 +0000 UTC m=+3536.974175555" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.489216 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.595780 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-catalog-content\") pod \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.596104 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-utilities\") pod \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.596201 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5tpd\" (UniqueName: \"kubernetes.io/projected/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-kube-api-access-t5tpd\") pod \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\" (UID: \"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f\") " Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.597091 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-utilities" (OuterVolumeSpecName: "utilities") pod "65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" (UID: "65f7c58a-f5a3-4c47-a7c2-c50f9de3395f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.615959 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-kube-api-access-t5tpd" (OuterVolumeSpecName: "kube-api-access-t5tpd") pod "65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" (UID: "65f7c58a-f5a3-4c47-a7c2-c50f9de3395f"). InnerVolumeSpecName "kube-api-access-t5tpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.647139 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" (UID: "65f7c58a-f5a3-4c47-a7c2-c50f9de3395f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.699362 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.699405 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.699418 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5tpd\" (UniqueName: \"kubernetes.io/projected/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f-kube-api-access-t5tpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.972652 4690 generic.go:334] "Generic (PLEG): container finished" podID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerID="e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b" exitCode=0 Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.972723 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28nm9" Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.972758 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28nm9" event={"ID":"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f","Type":"ContainerDied","Data":"e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b"} Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.974468 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28nm9" event={"ID":"65f7c58a-f5a3-4c47-a7c2-c50f9de3395f","Type":"ContainerDied","Data":"cc5294adc15597559ecd3d45b9b02a5326d890a2731cab9a10f2e211fa4087c5"} Dec 03 14:08:11 crc kubenswrapper[4690]: I1203 14:08:11.974496 4690 scope.go:117] "RemoveContainer" containerID="e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.009632 4690 scope.go:117] "RemoveContainer" containerID="893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.018436 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-28nm9"] Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.033854 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-28nm9"] Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.041485 4690 scope.go:117] "RemoveContainer" containerID="3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.082837 4690 scope.go:117] "RemoveContainer" containerID="e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b" Dec 03 14:08:12 crc kubenswrapper[4690]: E1203 14:08:12.083425 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b\": container with ID starting with e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b not found: ID does not exist" containerID="e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.083470 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b"} err="failed to get container status \"e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b\": rpc error: code = NotFound desc = could not find container \"e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b\": container with ID starting with e4b4893ea4b88d4f8c54e12058b1e7ee8e87bbb1f2c863fdd16f0c35812afd7b not found: ID does not exist" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.083493 4690 scope.go:117] "RemoveContainer" containerID="893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49" Dec 03 14:08:12 crc kubenswrapper[4690]: E1203 14:08:12.083821 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49\": container with ID starting with 893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49 not found: ID does not exist" containerID="893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.083884 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49"} err="failed to get container status \"893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49\": rpc error: code = NotFound desc = could not find container \"893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49\": container with ID starting with 893a142273beef8e6e2a9f4b8e61bb2ad71fffd3d832ed82c03d5402d50b9d49 not found: ID does not exist" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.083916 4690 scope.go:117] "RemoveContainer" containerID="3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8" Dec 03 14:08:12 crc kubenswrapper[4690]: E1203 14:08:12.084237 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8\": container with ID starting with 3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8 not found: ID does not exist" containerID="3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.084266 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8"} err="failed to get container status \"3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8\": rpc error: code = NotFound desc = could not find container \"3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8\": container with ID starting with 3387c76f9bb1fdbc4f6b078038f8f4e9affb53874c27242478a6ad20297c33e8 not found: ID does not exist" Dec 03 14:08:12 crc kubenswrapper[4690]: I1203 14:08:12.331385 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" path="/var/lib/kubelet/pods/65f7c58a-f5a3-4c47-a7c2-c50f9de3395f/volumes" Dec 03 14:08:46 crc kubenswrapper[4690]: I1203 14:08:46.823544 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:08:46 crc kubenswrapper[4690]: I1203 14:08:46.824190 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.883450 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8r62h"] Dec 03 14:09:10 crc kubenswrapper[4690]: E1203 14:09:10.884904 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerName="extract-utilities" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.884923 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerName="extract-utilities" Dec 03 14:09:10 crc kubenswrapper[4690]: E1203 14:09:10.884959 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerName="registry-server" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.884967 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerName="registry-server" Dec 03 14:09:10 crc kubenswrapper[4690]: E1203 14:09:10.884981 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerName="extract-content" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.884989 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerName="extract-content" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.885302 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f7c58a-f5a3-4c47-a7c2-c50f9de3395f" containerName="registry-server" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.895978 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.901149 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8r62h"] Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.982415 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhgmz\" (UniqueName: \"kubernetes.io/projected/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-kube-api-access-dhgmz\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.982598 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-utilities\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:10 crc kubenswrapper[4690]: I1203 14:09:10.982718 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-catalog-content\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:11 crc kubenswrapper[4690]: I1203 14:09:11.084787 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-utilities\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:11 crc kubenswrapper[4690]: I1203 14:09:11.084912 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-catalog-content\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:11 crc kubenswrapper[4690]: I1203 14:09:11.085000 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhgmz\" (UniqueName: \"kubernetes.io/projected/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-kube-api-access-dhgmz\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:11 crc kubenswrapper[4690]: I1203 14:09:11.085583 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-utilities\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:11 crc kubenswrapper[4690]: I1203 14:09:11.085615 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-catalog-content\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:11 crc kubenswrapper[4690]: I1203 14:09:11.108030 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhgmz\" (UniqueName: \"kubernetes.io/projected/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-kube-api-access-dhgmz\") pod \"redhat-marketplace-8r62h\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:11 crc kubenswrapper[4690]: I1203 14:09:11.229063 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:11 crc kubenswrapper[4690]: I1203 14:09:11.938992 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8r62h"] Dec 03 14:09:12 crc kubenswrapper[4690]: I1203 14:09:12.773296 4690 generic.go:334] "Generic (PLEG): container finished" podID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerID="68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2" exitCode=0 Dec 03 14:09:12 crc kubenswrapper[4690]: I1203 14:09:12.773347 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8r62h" event={"ID":"e76ec65c-3ea9-47e5-a396-98a1ee9584f2","Type":"ContainerDied","Data":"68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2"} Dec 03 14:09:12 crc kubenswrapper[4690]: I1203 14:09:12.773693 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8r62h" event={"ID":"e76ec65c-3ea9-47e5-a396-98a1ee9584f2","Type":"ContainerStarted","Data":"d86a4a9a6bacd850a3dbf216c1950dd6ec769832c4a586dfac86129f8fa36a9e"} Dec 03 14:09:14 crc kubenswrapper[4690]: I1203 14:09:14.820143 4690 generic.go:334] "Generic (PLEG): container finished" podID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerID="b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30" exitCode=0 Dec 03 14:09:14 crc kubenswrapper[4690]: I1203 14:09:14.820502 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8r62h" event={"ID":"e76ec65c-3ea9-47e5-a396-98a1ee9584f2","Type":"ContainerDied","Data":"b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30"} Dec 03 14:09:15 crc kubenswrapper[4690]: I1203 14:09:15.836794 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8r62h" event={"ID":"e76ec65c-3ea9-47e5-a396-98a1ee9584f2","Type":"ContainerStarted","Data":"d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76"} Dec 03 14:09:15 crc kubenswrapper[4690]: I1203 14:09:15.865739 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8r62h" podStartSLOduration=3.393474702 podStartE2EDuration="5.86570754s" podCreationTimestamp="2025-12-03 14:09:10 +0000 UTC" firstStartedPulling="2025-12-03 14:09:12.775508954 +0000 UTC m=+3598.756429387" lastFinishedPulling="2025-12-03 14:09:15.247741792 +0000 UTC m=+3601.228662225" observedRunningTime="2025-12-03 14:09:15.853900523 +0000 UTC m=+3601.834820966" watchObservedRunningTime="2025-12-03 14:09:15.86570754 +0000 UTC m=+3601.846627973" Dec 03 14:09:16 crc kubenswrapper[4690]: I1203 14:09:16.823924 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:09:16 crc kubenswrapper[4690]: I1203 14:09:16.824373 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:09:21 crc kubenswrapper[4690]: I1203 14:09:21.229606 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:21 crc kubenswrapper[4690]: I1203 14:09:21.230430 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:21 crc kubenswrapper[4690]: I1203 14:09:21.297147 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:21 crc kubenswrapper[4690]: I1203 14:09:21.985769 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:22 crc kubenswrapper[4690]: I1203 14:09:22.049049 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8r62h"] Dec 03 14:09:23 crc kubenswrapper[4690]: I1203 14:09:23.950013 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8r62h" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerName="registry-server" containerID="cri-o://d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76" gracePeriod=2 Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.665822 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.847929 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-catalog-content\") pod \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.848468 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhgmz\" (UniqueName: \"kubernetes.io/projected/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-kube-api-access-dhgmz\") pod \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.848610 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-utilities\") pod \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\" (UID: \"e76ec65c-3ea9-47e5-a396-98a1ee9584f2\") " Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.850341 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-utilities" (OuterVolumeSpecName: "utilities") pod "e76ec65c-3ea9-47e5-a396-98a1ee9584f2" (UID: "e76ec65c-3ea9-47e5-a396-98a1ee9584f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.886241 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-kube-api-access-dhgmz" (OuterVolumeSpecName: "kube-api-access-dhgmz") pod "e76ec65c-3ea9-47e5-a396-98a1ee9584f2" (UID: "e76ec65c-3ea9-47e5-a396-98a1ee9584f2"). InnerVolumeSpecName "kube-api-access-dhgmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.900611 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e76ec65c-3ea9-47e5-a396-98a1ee9584f2" (UID: "e76ec65c-3ea9-47e5-a396-98a1ee9584f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.951704 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhgmz\" (UniqueName: \"kubernetes.io/projected/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-kube-api-access-dhgmz\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.951743 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.951753 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76ec65c-3ea9-47e5-a396-98a1ee9584f2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.964148 4690 generic.go:334] "Generic (PLEG): container finished" podID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerID="d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76" exitCode=0 Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.964205 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8r62h" event={"ID":"e76ec65c-3ea9-47e5-a396-98a1ee9584f2","Type":"ContainerDied","Data":"d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76"} Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.964239 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8r62h" event={"ID":"e76ec65c-3ea9-47e5-a396-98a1ee9584f2","Type":"ContainerDied","Data":"d86a4a9a6bacd850a3dbf216c1950dd6ec769832c4a586dfac86129f8fa36a9e"} Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.964260 4690 scope.go:117] "RemoveContainer" containerID="d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76" Dec 03 14:09:24 crc kubenswrapper[4690]: I1203 14:09:24.964435 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8r62h" Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.007909 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8r62h"] Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.011583 4690 scope.go:117] "RemoveContainer" containerID="b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30" Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.029423 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8r62h"] Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.038158 4690 scope.go:117] "RemoveContainer" containerID="68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2" Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.093784 4690 scope.go:117] "RemoveContainer" containerID="d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76" Dec 03 14:09:25 crc kubenswrapper[4690]: E1203 14:09:25.094343 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76\": container with ID starting with d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76 not found: ID does not exist" containerID="d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76" Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.094374 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76"} err="failed to get container status \"d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76\": rpc error: code = NotFound desc = could not find container \"d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76\": container with ID starting with d29e5447ef2b3059719049fa80931289e4e99b084e897d9455f33e21f9765f76 not found: ID does not exist" Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.094407 4690 scope.go:117] "RemoveContainer" containerID="b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30" Dec 03 14:09:25 crc kubenswrapper[4690]: E1203 14:09:25.094738 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30\": container with ID starting with b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30 not found: ID does not exist" containerID="b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30" Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.094788 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30"} err="failed to get container status \"b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30\": rpc error: code = NotFound desc = could not find container \"b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30\": container with ID starting with b428d14df6e1d84492faad22bd6fa77699814af825b3631819d6ba6426f9be30 not found: ID does not exist" Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.094830 4690 scope.go:117] "RemoveContainer" containerID="68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2" Dec 03 14:09:25 crc kubenswrapper[4690]: E1203 14:09:25.095195 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2\": container with ID starting with 68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2 not found: ID does not exist" containerID="68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2" Dec 03 14:09:25 crc kubenswrapper[4690]: I1203 14:09:25.095219 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2"} err="failed to get container status \"68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2\": rpc error: code = NotFound desc = could not find container \"68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2\": container with ID starting with 68293a47a73025c7c250f8030d877acf5bcc75f6c0074ee97ec582bd395f9bb2 not found: ID does not exist" Dec 03 14:09:26 crc kubenswrapper[4690]: I1203 14:09:26.328419 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" path="/var/lib/kubelet/pods/e76ec65c-3ea9-47e5-a396-98a1ee9584f2/volumes" Dec 03 14:09:46 crc kubenswrapper[4690]: I1203 14:09:46.824050 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:09:46 crc kubenswrapper[4690]: I1203 14:09:46.824706 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:09:46 crc kubenswrapper[4690]: I1203 14:09:46.824755 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 14:09:46 crc kubenswrapper[4690]: I1203 14:09:46.825674 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c88144a8ab40e991dd2e75e0df62f1bf2a23d73e39a209a9eec846dd271c877"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:09:46 crc kubenswrapper[4690]: I1203 14:09:46.825740 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://0c88144a8ab40e991dd2e75e0df62f1bf2a23d73e39a209a9eec846dd271c877" gracePeriod=600 Dec 03 14:09:47 crc kubenswrapper[4690]: I1203 14:09:47.199038 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="0c88144a8ab40e991dd2e75e0df62f1bf2a23d73e39a209a9eec846dd271c877" exitCode=0 Dec 03 14:09:47 crc kubenswrapper[4690]: I1203 14:09:47.199124 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"0c88144a8ab40e991dd2e75e0df62f1bf2a23d73e39a209a9eec846dd271c877"} Dec 03 14:09:47 crc kubenswrapper[4690]: I1203 14:09:47.199525 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af"} Dec 03 14:09:47 crc kubenswrapper[4690]: I1203 14:09:47.199551 4690 scope.go:117] "RemoveContainer" containerID="ec3522b1146c859a7ae4176a384210c062378c1a338cfe7a9c7d145632bc6615" Dec 03 14:12:16 crc kubenswrapper[4690]: I1203 14:12:16.823717 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:12:16 crc kubenswrapper[4690]: I1203 14:12:16.824421 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:12:46 crc kubenswrapper[4690]: I1203 14:12:46.823432 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:12:46 crc kubenswrapper[4690]: I1203 14:12:46.824197 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:13:03 crc kubenswrapper[4690]: I1203 14:13:03.278036 4690 generic.go:334] "Generic (PLEG): container finished" podID="46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" containerID="d01735d62d236c5097c74d0a792ce3463f51495439f7e412e1dd93f87ebfb4d2" exitCode=0 Dec 03 14:13:03 crc kubenswrapper[4690]: I1203 14:13:03.278137 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5","Type":"ContainerDied","Data":"d01735d62d236c5097c74d0a792ce3463f51495439f7e412e1dd93f87ebfb4d2"} Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.854529 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965429 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-temporary\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965492 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965593 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-config-data\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965645 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ssh-key\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965665 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ca-certs\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965710 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8f6w\" (UniqueName: \"kubernetes.io/projected/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-kube-api-access-r8f6w\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965837 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-workdir\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965913 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config-secret\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.965975 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config\") pod \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\" (UID: \"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5\") " Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.966233 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.966366 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-config-data" (OuterVolumeSpecName: "config-data") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.966684 4690 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.966712 4690 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.970345 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.972185 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.973168 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-kube-api-access-r8f6w" (OuterVolumeSpecName: "kube-api-access-r8f6w") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "kube-api-access-r8f6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:04 crc kubenswrapper[4690]: I1203 14:13:04.998586 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.000466 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.003291 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.025162 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" (UID: "46b67a35-a83b-48dd-a6ea-4464ff4ee3d5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.069092 4690 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.069145 4690 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.069156 4690 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.069172 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8f6w\" (UniqueName: \"kubernetes.io/projected/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-kube-api-access-r8f6w\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.069186 4690 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.069199 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.069211 4690 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/46b67a35-a83b-48dd-a6ea-4464ff4ee3d5-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.091674 4690 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.170998 4690 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.302052 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"46b67a35-a83b-48dd-a6ea-4464ff4ee3d5","Type":"ContainerDied","Data":"859b71dcf00cbc746d37fc58876e5a0339190023821a9c978bb6b424f9d2d944"} Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.302105 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="859b71dcf00cbc746d37fc58876e5a0339190023821a9c978bb6b424f9d2d944" Dec 03 14:13:05 crc kubenswrapper[4690]: I1203 14:13:05.302134 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.644034 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 14:13:11 crc kubenswrapper[4690]: E1203 14:13:11.645315 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerName="extract-utilities" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.645340 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerName="extract-utilities" Dec 03 14:13:11 crc kubenswrapper[4690]: E1203 14:13:11.645363 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" containerName="tempest-tests-tempest-tests-runner" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.645376 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" containerName="tempest-tests-tempest-tests-runner" Dec 03 14:13:11 crc kubenswrapper[4690]: E1203 14:13:11.645404 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerName="extract-content" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.645412 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerName="extract-content" Dec 03 14:13:11 crc kubenswrapper[4690]: E1203 14:13:11.645432 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerName="registry-server" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.645440 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerName="registry-server" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.645774 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b67a35-a83b-48dd-a6ea-4464ff4ee3d5" containerName="tempest-tests-tempest-tests-runner" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.645798 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76ec65c-3ea9-47e5-a396-98a1ee9584f2" containerName="registry-server" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.646767 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.649797 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cbk7k" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.656205 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.812778 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7b33ef02-9b68-41ad-90ed-1adfe0cd8312\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.812847 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqt9f\" (UniqueName: \"kubernetes.io/projected/7b33ef02-9b68-41ad-90ed-1adfe0cd8312-kube-api-access-kqt9f\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7b33ef02-9b68-41ad-90ed-1adfe0cd8312\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.915708 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7b33ef02-9b68-41ad-90ed-1adfe0cd8312\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.915798 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqt9f\" (UniqueName: \"kubernetes.io/projected/7b33ef02-9b68-41ad-90ed-1adfe0cd8312-kube-api-access-kqt9f\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7b33ef02-9b68-41ad-90ed-1adfe0cd8312\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.916325 4690 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7b33ef02-9b68-41ad-90ed-1adfe0cd8312\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.955652 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqt9f\" (UniqueName: \"kubernetes.io/projected/7b33ef02-9b68-41ad-90ed-1adfe0cd8312-kube-api-access-kqt9f\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7b33ef02-9b68-41ad-90ed-1adfe0cd8312\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:11 crc kubenswrapper[4690]: I1203 14:13:11.995030 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7b33ef02-9b68-41ad-90ed-1adfe0cd8312\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:12 crc kubenswrapper[4690]: I1203 14:13:12.273466 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 14:13:12 crc kubenswrapper[4690]: I1203 14:13:12.831103 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 14:13:12 crc kubenswrapper[4690]: I1203 14:13:12.837391 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:13:13 crc kubenswrapper[4690]: I1203 14:13:13.394822 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"7b33ef02-9b68-41ad-90ed-1adfe0cd8312","Type":"ContainerStarted","Data":"a4aa1de6be0c9f9fd9c7975bbc987b7037b85fb19f1da548b20b1efdb8fc54fb"} Dec 03 14:13:15 crc kubenswrapper[4690]: I1203 14:13:15.424445 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"7b33ef02-9b68-41ad-90ed-1adfe0cd8312","Type":"ContainerStarted","Data":"f0b200bbda65e11edefa15ceaf31a04f5762889cb6951725c291baa55a8d6c48"} Dec 03 14:13:15 crc kubenswrapper[4690]: I1203 14:13:15.445858 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.138528228 podStartE2EDuration="4.445828893s" podCreationTimestamp="2025-12-03 14:13:11 +0000 UTC" firstStartedPulling="2025-12-03 14:13:12.837130931 +0000 UTC m=+3838.818051364" lastFinishedPulling="2025-12-03 14:13:15.144431596 +0000 UTC m=+3841.125352029" observedRunningTime="2025-12-03 14:13:15.436070597 +0000 UTC m=+3841.416991030" watchObservedRunningTime="2025-12-03 14:13:15.445828893 +0000 UTC m=+3841.426749336" Dec 03 14:13:16 crc kubenswrapper[4690]: I1203 14:13:16.823459 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:13:16 crc kubenswrapper[4690]: I1203 14:13:16.823863 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:13:16 crc kubenswrapper[4690]: I1203 14:13:16.823930 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 14:13:16 crc kubenswrapper[4690]: I1203 14:13:16.824784 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:13:16 crc kubenswrapper[4690]: I1203 14:13:16.824832 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" gracePeriod=600 Dec 03 14:13:16 crc kubenswrapper[4690]: E1203 14:13:16.986963 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:13:17 crc kubenswrapper[4690]: I1203 14:13:17.448697 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" exitCode=0 Dec 03 14:13:17 crc kubenswrapper[4690]: I1203 14:13:17.448785 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af"} Dec 03 14:13:17 crc kubenswrapper[4690]: I1203 14:13:17.449139 4690 scope.go:117] "RemoveContainer" containerID="0c88144a8ab40e991dd2e75e0df62f1bf2a23d73e39a209a9eec846dd271c877" Dec 03 14:13:17 crc kubenswrapper[4690]: I1203 14:13:17.450020 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:13:17 crc kubenswrapper[4690]: E1203 14:13:17.450340 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:13:31 crc kubenswrapper[4690]: I1203 14:13:31.315546 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:13:31 crc kubenswrapper[4690]: E1203 14:13:31.316608 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.300526 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-57qfs/must-gather-lzcsm"] Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.304993 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.310202 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-57qfs"/"openshift-service-ca.crt" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.310329 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-57qfs"/"default-dockercfg-pfg7g" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.310717 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-57qfs"/"kube-root-ca.crt" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.336391 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-57qfs/must-gather-lzcsm"] Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.439760 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf253854-f14d-4ea6-9f4b-6e988d5fb432-must-gather-output\") pod \"must-gather-lzcsm\" (UID: \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\") " pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.440255 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp5vh\" (UniqueName: \"kubernetes.io/projected/cf253854-f14d-4ea6-9f4b-6e988d5fb432-kube-api-access-lp5vh\") pod \"must-gather-lzcsm\" (UID: \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\") " pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.542563 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp5vh\" (UniqueName: \"kubernetes.io/projected/cf253854-f14d-4ea6-9f4b-6e988d5fb432-kube-api-access-lp5vh\") pod \"must-gather-lzcsm\" (UID: \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\") " pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.543055 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf253854-f14d-4ea6-9f4b-6e988d5fb432-must-gather-output\") pod \"must-gather-lzcsm\" (UID: \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\") " pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.543623 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf253854-f14d-4ea6-9f4b-6e988d5fb432-must-gather-output\") pod \"must-gather-lzcsm\" (UID: \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\") " pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.566106 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp5vh\" (UniqueName: \"kubernetes.io/projected/cf253854-f14d-4ea6-9f4b-6e988d5fb432-kube-api-access-lp5vh\") pod \"must-gather-lzcsm\" (UID: \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\") " pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:13:42 crc kubenswrapper[4690]: I1203 14:13:42.635101 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:13:43 crc kubenswrapper[4690]: I1203 14:13:43.297656 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-57qfs/must-gather-lzcsm"] Dec 03 14:13:43 crc kubenswrapper[4690]: I1203 14:13:43.778006 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/must-gather-lzcsm" event={"ID":"cf253854-f14d-4ea6-9f4b-6e988d5fb432","Type":"ContainerStarted","Data":"fbf0104d2d770f26ec912fc5e28d5de5d30ab458ba96e71153f7066ee77fc457"} Dec 03 14:13:46 crc kubenswrapper[4690]: I1203 14:13:46.318548 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:13:46 crc kubenswrapper[4690]: E1203 14:13:46.319503 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:13:48 crc kubenswrapper[4690]: I1203 14:13:48.832692 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/must-gather-lzcsm" event={"ID":"cf253854-f14d-4ea6-9f4b-6e988d5fb432","Type":"ContainerStarted","Data":"2d2b632a39484f16c71618a830f8238bbc2002337fe94ba6fef8f64b89a7c929"} Dec 03 14:13:48 crc kubenswrapper[4690]: I1203 14:13:48.833404 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/must-gather-lzcsm" event={"ID":"cf253854-f14d-4ea6-9f4b-6e988d5fb432","Type":"ContainerStarted","Data":"e343e8a0a0ab1358feb642c0c16e1dcc26de2fbbd7c0c36b563162cefec04d7b"} Dec 03 14:13:48 crc kubenswrapper[4690]: I1203 14:13:48.849508 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-57qfs/must-gather-lzcsm" podStartSLOduration=2.406350908 podStartE2EDuration="6.849481789s" podCreationTimestamp="2025-12-03 14:13:42 +0000 UTC" firstStartedPulling="2025-12-03 14:13:43.306648932 +0000 UTC m=+3869.287569375" lastFinishedPulling="2025-12-03 14:13:47.749779823 +0000 UTC m=+3873.730700256" observedRunningTime="2025-12-03 14:13:48.845393456 +0000 UTC m=+3874.826313889" watchObservedRunningTime="2025-12-03 14:13:48.849481789 +0000 UTC m=+3874.830402222" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.431852 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-57qfs/crc-debug-lqv68"] Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.435021 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.529658 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgz2n\" (UniqueName: \"kubernetes.io/projected/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-kube-api-access-zgz2n\") pod \"crc-debug-lqv68\" (UID: \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\") " pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.529941 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-host\") pod \"crc-debug-lqv68\" (UID: \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\") " pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.632183 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-host\") pod \"crc-debug-lqv68\" (UID: \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\") " pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.632263 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgz2n\" (UniqueName: \"kubernetes.io/projected/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-kube-api-access-zgz2n\") pod \"crc-debug-lqv68\" (UID: \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\") " pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.632299 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-host\") pod \"crc-debug-lqv68\" (UID: \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\") " pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.657735 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgz2n\" (UniqueName: \"kubernetes.io/projected/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-kube-api-access-zgz2n\") pod \"crc-debug-lqv68\" (UID: \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\") " pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.760334 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:13:54 crc kubenswrapper[4690]: I1203 14:13:54.910982 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/crc-debug-lqv68" event={"ID":"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e","Type":"ContainerStarted","Data":"b74fbcb843c5666d48cd820ed6145418173683f9f6730d4506669ae70b6d2496"} Dec 03 14:13:57 crc kubenswrapper[4690]: I1203 14:13:57.314497 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:13:57 crc kubenswrapper[4690]: E1203 14:13:57.315446 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:14:09 crc kubenswrapper[4690]: E1203 14:14:09.949278 4690 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Dec 03 14:14:09 crc kubenswrapper[4690]: E1203 14:14:09.950128 4690 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zgz2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-lqv68_openshift-must-gather-57qfs(aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:14:09 crc kubenswrapper[4690]: E1203 14:14:09.951627 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-57qfs/crc-debug-lqv68" podUID="aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" Dec 03 14:14:10 crc kubenswrapper[4690]: E1203 14:14:10.063655 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-57qfs/crc-debug-lqv68" podUID="aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" Dec 03 14:14:11 crc kubenswrapper[4690]: I1203 14:14:11.315445 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:14:11 crc kubenswrapper[4690]: E1203 14:14:11.316254 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:14:24 crc kubenswrapper[4690]: I1203 14:14:24.216250 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/crc-debug-lqv68" event={"ID":"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e","Type":"ContainerStarted","Data":"9d5fda47b9a1f1971367b4cd4ba19bd1f6aa17828a1bcc5179ee9ad6d9fc2cf3"} Dec 03 14:14:24 crc kubenswrapper[4690]: I1203 14:14:24.240224 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-57qfs/crc-debug-lqv68" podStartSLOduration=1.502918317 podStartE2EDuration="30.240197458s" podCreationTimestamp="2025-12-03 14:13:54 +0000 UTC" firstStartedPulling="2025-12-03 14:13:54.82036587 +0000 UTC m=+3880.801286303" lastFinishedPulling="2025-12-03 14:14:23.557645011 +0000 UTC m=+3909.538565444" observedRunningTime="2025-12-03 14:14:24.232362591 +0000 UTC m=+3910.213283024" watchObservedRunningTime="2025-12-03 14:14:24.240197458 +0000 UTC m=+3910.221117891" Dec 03 14:14:25 crc kubenswrapper[4690]: I1203 14:14:25.315999 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:14:25 crc kubenswrapper[4690]: E1203 14:14:25.316756 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:14:36 crc kubenswrapper[4690]: I1203 14:14:36.314690 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:14:36 crc kubenswrapper[4690]: E1203 14:14:36.316933 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:14:49 crc kubenswrapper[4690]: I1203 14:14:49.314991 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:14:49 crc kubenswrapper[4690]: E1203 14:14:49.318291 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.189090 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6"] Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.191713 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.196515 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.198789 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.215485 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6"] Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.305993 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b868aa93-206d-45c2-92f7-21bb4069ac41-secret-volume\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.306099 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b868aa93-206d-45c2-92f7-21bb4069ac41-config-volume\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.306180 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7swmg\" (UniqueName: \"kubernetes.io/projected/b868aa93-206d-45c2-92f7-21bb4069ac41-kube-api-access-7swmg\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.316240 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:15:00 crc kubenswrapper[4690]: E1203 14:15:00.316907 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.408538 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b868aa93-206d-45c2-92f7-21bb4069ac41-config-volume\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.409677 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7swmg\" (UniqueName: \"kubernetes.io/projected/b868aa93-206d-45c2-92f7-21bb4069ac41-kube-api-access-7swmg\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.409996 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b868aa93-206d-45c2-92f7-21bb4069ac41-config-volume\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.411328 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b868aa93-206d-45c2-92f7-21bb4069ac41-secret-volume\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.418153 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b868aa93-206d-45c2-92f7-21bb4069ac41-secret-volume\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.429305 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7swmg\" (UniqueName: \"kubernetes.io/projected/b868aa93-206d-45c2-92f7-21bb4069ac41-kube-api-access-7swmg\") pod \"collect-profiles-29412855-6p5s6\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:00 crc kubenswrapper[4690]: I1203 14:15:00.516104 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:01 crc kubenswrapper[4690]: I1203 14:15:01.013066 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6"] Dec 03 14:15:01 crc kubenswrapper[4690]: I1203 14:15:01.620707 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" event={"ID":"b868aa93-206d-45c2-92f7-21bb4069ac41","Type":"ContainerStarted","Data":"a093ae02279d990fdd8ae2b8ba47fb961273c9391e97b2fbf8f611418effdac1"} Dec 03 14:15:02 crc kubenswrapper[4690]: I1203 14:15:02.634531 4690 generic.go:334] "Generic (PLEG): container finished" podID="b868aa93-206d-45c2-92f7-21bb4069ac41" containerID="4e9abb94cd9376976c0166ac4fe68a7668c50f5be581d52c533df580e20c3f41" exitCode=0 Dec 03 14:15:02 crc kubenswrapper[4690]: I1203 14:15:02.634645 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" event={"ID":"b868aa93-206d-45c2-92f7-21bb4069ac41","Type":"ContainerDied","Data":"4e9abb94cd9376976c0166ac4fe68a7668c50f5be581d52c533df580e20c3f41"} Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.091696 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.198816 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b868aa93-206d-45c2-92f7-21bb4069ac41-config-volume\") pod \"b868aa93-206d-45c2-92f7-21bb4069ac41\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.199010 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7swmg\" (UniqueName: \"kubernetes.io/projected/b868aa93-206d-45c2-92f7-21bb4069ac41-kube-api-access-7swmg\") pod \"b868aa93-206d-45c2-92f7-21bb4069ac41\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.199040 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b868aa93-206d-45c2-92f7-21bb4069ac41-secret-volume\") pod \"b868aa93-206d-45c2-92f7-21bb4069ac41\" (UID: \"b868aa93-206d-45c2-92f7-21bb4069ac41\") " Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.199783 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b868aa93-206d-45c2-92f7-21bb4069ac41-config-volume" (OuterVolumeSpecName: "config-volume") pod "b868aa93-206d-45c2-92f7-21bb4069ac41" (UID: "b868aa93-206d-45c2-92f7-21bb4069ac41"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.203368 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b868aa93-206d-45c2-92f7-21bb4069ac41-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.206411 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b868aa93-206d-45c2-92f7-21bb4069ac41-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b868aa93-206d-45c2-92f7-21bb4069ac41" (UID: "b868aa93-206d-45c2-92f7-21bb4069ac41"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.206691 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b868aa93-206d-45c2-92f7-21bb4069ac41-kube-api-access-7swmg" (OuterVolumeSpecName: "kube-api-access-7swmg") pod "b868aa93-206d-45c2-92f7-21bb4069ac41" (UID: "b868aa93-206d-45c2-92f7-21bb4069ac41"). InnerVolumeSpecName "kube-api-access-7swmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.305473 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7swmg\" (UniqueName: \"kubernetes.io/projected/b868aa93-206d-45c2-92f7-21bb4069ac41-kube-api-access-7swmg\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.305520 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b868aa93-206d-45c2-92f7-21bb4069ac41-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.654445 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" event={"ID":"b868aa93-206d-45c2-92f7-21bb4069ac41","Type":"ContainerDied","Data":"a093ae02279d990fdd8ae2b8ba47fb961273c9391e97b2fbf8f611418effdac1"} Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.654814 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a093ae02279d990fdd8ae2b8ba47fb961273c9391e97b2fbf8f611418effdac1" Dec 03 14:15:04 crc kubenswrapper[4690]: I1203 14:15:04.655041 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-6p5s6" Dec 03 14:15:05 crc kubenswrapper[4690]: I1203 14:15:05.180991 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2"] Dec 03 14:15:05 crc kubenswrapper[4690]: I1203 14:15:05.196623 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412810-x2cl2"] Dec 03 14:15:06 crc kubenswrapper[4690]: I1203 14:15:06.332639 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9edbc53-d827-45be-aa22-bb691422fb4d" path="/var/lib/kubelet/pods/d9edbc53-d827-45be-aa22-bb691422fb4d/volumes" Dec 03 14:15:14 crc kubenswrapper[4690]: I1203 14:15:14.324524 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:15:14 crc kubenswrapper[4690]: E1203 14:15:14.325500 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:15:14 crc kubenswrapper[4690]: I1203 14:15:14.762256 4690 generic.go:334] "Generic (PLEG): container finished" podID="aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" containerID="9d5fda47b9a1f1971367b4cd4ba19bd1f6aa17828a1bcc5179ee9ad6d9fc2cf3" exitCode=0 Dec 03 14:15:14 crc kubenswrapper[4690]: I1203 14:15:14.762361 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/crc-debug-lqv68" event={"ID":"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e","Type":"ContainerDied","Data":"9d5fda47b9a1f1971367b4cd4ba19bd1f6aa17828a1bcc5179ee9ad6d9fc2cf3"} Dec 03 14:15:15 crc kubenswrapper[4690]: I1203 14:15:15.885950 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:15:15 crc kubenswrapper[4690]: I1203 14:15:15.921678 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-57qfs/crc-debug-lqv68"] Dec 03 14:15:15 crc kubenswrapper[4690]: I1203 14:15:15.931250 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-57qfs/crc-debug-lqv68"] Dec 03 14:15:15 crc kubenswrapper[4690]: I1203 14:15:15.964969 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-host\") pod \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\" (UID: \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\") " Dec 03 14:15:15 crc kubenswrapper[4690]: I1203 14:15:15.965049 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgz2n\" (UniqueName: \"kubernetes.io/projected/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-kube-api-access-zgz2n\") pod \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\" (UID: \"aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e\") " Dec 03 14:15:15 crc kubenswrapper[4690]: I1203 14:15:15.965092 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-host" (OuterVolumeSpecName: "host") pod "aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" (UID: "aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:15:15 crc kubenswrapper[4690]: I1203 14:15:15.965772 4690 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-host\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:15 crc kubenswrapper[4690]: I1203 14:15:15.971234 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-kube-api-access-zgz2n" (OuterVolumeSpecName: "kube-api-access-zgz2n") pod "aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" (UID: "aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e"). InnerVolumeSpecName "kube-api-access-zgz2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4690]: I1203 14:15:16.068277 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgz2n\" (UniqueName: \"kubernetes.io/projected/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e-kube-api-access-zgz2n\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:16 crc kubenswrapper[4690]: I1203 14:15:16.330577 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" path="/var/lib/kubelet/pods/aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e/volumes" Dec 03 14:15:16 crc kubenswrapper[4690]: I1203 14:15:16.785546 4690 scope.go:117] "RemoveContainer" containerID="9d5fda47b9a1f1971367b4cd4ba19bd1f6aa17828a1bcc5179ee9ad6d9fc2cf3" Dec 03 14:15:16 crc kubenswrapper[4690]: I1203 14:15:16.785786 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-lqv68" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.099584 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-57qfs/crc-debug-pqqz8"] Dec 03 14:15:17 crc kubenswrapper[4690]: E1203 14:15:17.100526 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b868aa93-206d-45c2-92f7-21bb4069ac41" containerName="collect-profiles" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.100542 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b868aa93-206d-45c2-92f7-21bb4069ac41" containerName="collect-profiles" Dec 03 14:15:17 crc kubenswrapper[4690]: E1203 14:15:17.100583 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" containerName="container-00" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.100589 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" containerName="container-00" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.100808 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa6b1e3-dfbf-4754-8785-9bb1ceb2c66e" containerName="container-00" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.100824 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b868aa93-206d-45c2-92f7-21bb4069ac41" containerName="collect-profiles" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.101537 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.192476 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m9hr\" (UniqueName: \"kubernetes.io/projected/e713586c-b246-42b9-b8d5-59d9002e02b0-kube-api-access-8m9hr\") pod \"crc-debug-pqqz8\" (UID: \"e713586c-b246-42b9-b8d5-59d9002e02b0\") " pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.192799 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e713586c-b246-42b9-b8d5-59d9002e02b0-host\") pod \"crc-debug-pqqz8\" (UID: \"e713586c-b246-42b9-b8d5-59d9002e02b0\") " pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.295474 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m9hr\" (UniqueName: \"kubernetes.io/projected/e713586c-b246-42b9-b8d5-59d9002e02b0-kube-api-access-8m9hr\") pod \"crc-debug-pqqz8\" (UID: \"e713586c-b246-42b9-b8d5-59d9002e02b0\") " pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.295555 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e713586c-b246-42b9-b8d5-59d9002e02b0-host\") pod \"crc-debug-pqqz8\" (UID: \"e713586c-b246-42b9-b8d5-59d9002e02b0\") " pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.295773 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e713586c-b246-42b9-b8d5-59d9002e02b0-host\") pod \"crc-debug-pqqz8\" (UID: \"e713586c-b246-42b9-b8d5-59d9002e02b0\") " pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.318620 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m9hr\" (UniqueName: \"kubernetes.io/projected/e713586c-b246-42b9-b8d5-59d9002e02b0-kube-api-access-8m9hr\") pod \"crc-debug-pqqz8\" (UID: \"e713586c-b246-42b9-b8d5-59d9002e02b0\") " pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.421464 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:17 crc kubenswrapper[4690]: I1203 14:15:17.796733 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/crc-debug-pqqz8" event={"ID":"e713586c-b246-42b9-b8d5-59d9002e02b0","Type":"ContainerStarted","Data":"749be7380949a9f064e06ef57b899e1088eead99cdcc33a412df21076e8c7e54"} Dec 03 14:15:18 crc kubenswrapper[4690]: I1203 14:15:18.814135 4690 generic.go:334] "Generic (PLEG): container finished" podID="e713586c-b246-42b9-b8d5-59d9002e02b0" containerID="af4264b47017319e1b795f14ae07bd8d233bff9942436ae215d3bd0a27c3e6df" exitCode=0 Dec 03 14:15:18 crc kubenswrapper[4690]: I1203 14:15:18.814329 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/crc-debug-pqqz8" event={"ID":"e713586c-b246-42b9-b8d5-59d9002e02b0","Type":"ContainerDied","Data":"af4264b47017319e1b795f14ae07bd8d233bff9942436ae215d3bd0a27c3e6df"} Dec 03 14:15:19 crc kubenswrapper[4690]: I1203 14:15:19.946691 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.055718 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-57qfs/crc-debug-pqqz8"] Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.059730 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e713586c-b246-42b9-b8d5-59d9002e02b0-host\") pod \"e713586c-b246-42b9-b8d5-59d9002e02b0\" (UID: \"e713586c-b246-42b9-b8d5-59d9002e02b0\") " Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.059834 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m9hr\" (UniqueName: \"kubernetes.io/projected/e713586c-b246-42b9-b8d5-59d9002e02b0-kube-api-access-8m9hr\") pod \"e713586c-b246-42b9-b8d5-59d9002e02b0\" (UID: \"e713586c-b246-42b9-b8d5-59d9002e02b0\") " Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.059919 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e713586c-b246-42b9-b8d5-59d9002e02b0-host" (OuterVolumeSpecName: "host") pod "e713586c-b246-42b9-b8d5-59d9002e02b0" (UID: "e713586c-b246-42b9-b8d5-59d9002e02b0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.060556 4690 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e713586c-b246-42b9-b8d5-59d9002e02b0-host\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.066696 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e713586c-b246-42b9-b8d5-59d9002e02b0-kube-api-access-8m9hr" (OuterVolumeSpecName: "kube-api-access-8m9hr") pod "e713586c-b246-42b9-b8d5-59d9002e02b0" (UID: "e713586c-b246-42b9-b8d5-59d9002e02b0"). InnerVolumeSpecName "kube-api-access-8m9hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.070565 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-57qfs/crc-debug-pqqz8"] Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.162786 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m9hr\" (UniqueName: \"kubernetes.io/projected/e713586c-b246-42b9-b8d5-59d9002e02b0-kube-api-access-8m9hr\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.328646 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e713586c-b246-42b9-b8d5-59d9002e02b0" path="/var/lib/kubelet/pods/e713586c-b246-42b9-b8d5-59d9002e02b0/volumes" Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.836745 4690 scope.go:117] "RemoveContainer" containerID="af4264b47017319e1b795f14ae07bd8d233bff9942436ae215d3bd0a27c3e6df" Dec 03 14:15:20 crc kubenswrapper[4690]: I1203 14:15:20.836836 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-pqqz8" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.368062 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-57qfs/crc-debug-smv47"] Dec 03 14:15:21 crc kubenswrapper[4690]: E1203 14:15:21.368626 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e713586c-b246-42b9-b8d5-59d9002e02b0" containerName="container-00" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.368644 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="e713586c-b246-42b9-b8d5-59d9002e02b0" containerName="container-00" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.368883 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="e713586c-b246-42b9-b8d5-59d9002e02b0" containerName="container-00" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.369843 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.397888 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5521c45e-10ae-465e-8e25-628e67182be1-host\") pod \"crc-debug-smv47\" (UID: \"5521c45e-10ae-465e-8e25-628e67182be1\") " pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.397996 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlz69\" (UniqueName: \"kubernetes.io/projected/5521c45e-10ae-465e-8e25-628e67182be1-kube-api-access-jlz69\") pod \"crc-debug-smv47\" (UID: \"5521c45e-10ae-465e-8e25-628e67182be1\") " pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.500819 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5521c45e-10ae-465e-8e25-628e67182be1-host\") pod \"crc-debug-smv47\" (UID: \"5521c45e-10ae-465e-8e25-628e67182be1\") " pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.500927 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlz69\" (UniqueName: \"kubernetes.io/projected/5521c45e-10ae-465e-8e25-628e67182be1-kube-api-access-jlz69\") pod \"crc-debug-smv47\" (UID: \"5521c45e-10ae-465e-8e25-628e67182be1\") " pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.501363 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5521c45e-10ae-465e-8e25-628e67182be1-host\") pod \"crc-debug-smv47\" (UID: \"5521c45e-10ae-465e-8e25-628e67182be1\") " pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.531615 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlz69\" (UniqueName: \"kubernetes.io/projected/5521c45e-10ae-465e-8e25-628e67182be1-kube-api-access-jlz69\") pod \"crc-debug-smv47\" (UID: \"5521c45e-10ae-465e-8e25-628e67182be1\") " pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.688947 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:21 crc kubenswrapper[4690]: I1203 14:15:21.853347 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/crc-debug-smv47" event={"ID":"5521c45e-10ae-465e-8e25-628e67182be1","Type":"ContainerStarted","Data":"64585d6f19c061f49b656a0e56c2f5eee218e68986aea442c5326ed359aa3266"} Dec 03 14:15:22 crc kubenswrapper[4690]: I1203 14:15:22.865673 4690 generic.go:334] "Generic (PLEG): container finished" podID="5521c45e-10ae-465e-8e25-628e67182be1" containerID="9f3d68c50d5b67afc0455fba939a81c580380ed4c86e845f264dc6a50734f28a" exitCode=0 Dec 03 14:15:22 crc kubenswrapper[4690]: I1203 14:15:22.865740 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/crc-debug-smv47" event={"ID":"5521c45e-10ae-465e-8e25-628e67182be1","Type":"ContainerDied","Data":"9f3d68c50d5b67afc0455fba939a81c580380ed4c86e845f264dc6a50734f28a"} Dec 03 14:15:22 crc kubenswrapper[4690]: I1203 14:15:22.911640 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-57qfs/crc-debug-smv47"] Dec 03 14:15:22 crc kubenswrapper[4690]: I1203 14:15:22.921162 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-57qfs/crc-debug-smv47"] Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.018793 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.061732 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5521c45e-10ae-465e-8e25-628e67182be1-host\") pod \"5521c45e-10ae-465e-8e25-628e67182be1\" (UID: \"5521c45e-10ae-465e-8e25-628e67182be1\") " Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.061933 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5521c45e-10ae-465e-8e25-628e67182be1-host" (OuterVolumeSpecName: "host") pod "5521c45e-10ae-465e-8e25-628e67182be1" (UID: "5521c45e-10ae-465e-8e25-628e67182be1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.062231 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlz69\" (UniqueName: \"kubernetes.io/projected/5521c45e-10ae-465e-8e25-628e67182be1-kube-api-access-jlz69\") pod \"5521c45e-10ae-465e-8e25-628e67182be1\" (UID: \"5521c45e-10ae-465e-8e25-628e67182be1\") " Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.062920 4690 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5521c45e-10ae-465e-8e25-628e67182be1-host\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.089367 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5521c45e-10ae-465e-8e25-628e67182be1-kube-api-access-jlz69" (OuterVolumeSpecName: "kube-api-access-jlz69") pod "5521c45e-10ae-465e-8e25-628e67182be1" (UID: "5521c45e-10ae-465e-8e25-628e67182be1"). InnerVolumeSpecName "kube-api-access-jlz69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.165654 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlz69\" (UniqueName: \"kubernetes.io/projected/5521c45e-10ae-465e-8e25-628e67182be1-kube-api-access-jlz69\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.329446 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5521c45e-10ae-465e-8e25-628e67182be1" path="/var/lib/kubelet/pods/5521c45e-10ae-465e-8e25-628e67182be1/volumes" Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.892109 4690 scope.go:117] "RemoveContainer" containerID="9f3d68c50d5b67afc0455fba939a81c580380ed4c86e845f264dc6a50734f28a" Dec 03 14:15:24 crc kubenswrapper[4690]: I1203 14:15:24.892187 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/crc-debug-smv47" Dec 03 14:15:27 crc kubenswrapper[4690]: I1203 14:15:27.315485 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:15:27 crc kubenswrapper[4690]: E1203 14:15:27.316399 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:15:38 crc kubenswrapper[4690]: I1203 14:15:38.321471 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:15:38 crc kubenswrapper[4690]: E1203 14:15:38.322703 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:15:42 crc kubenswrapper[4690]: I1203 14:15:42.749453 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d8d5ff786-rbh82_1a74f0f7-bcc8-483f-ad5a-2df09e7f462d/barbican-api/0.log" Dec 03 14:15:42 crc kubenswrapper[4690]: I1203 14:15:42.925257 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d8d5ff786-rbh82_1a74f0f7-bcc8-483f-ad5a-2df09e7f462d/barbican-api-log/0.log" Dec 03 14:15:42 crc kubenswrapper[4690]: I1203 14:15:42.931127 4690 scope.go:117] "RemoveContainer" containerID="227845436ebc018f13d5937233cfc10d0acbf689be325a7635063e7be03af4cd" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.029879 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-98c9dfb78-qjztd_790eb625-6e7d-4467-afef-4073973889b2/barbican-keystone-listener/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.184239 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-98c9dfb78-qjztd_790eb625-6e7d-4467-afef-4073973889b2/barbican-keystone-listener-log/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.280812 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7c7c4b9c77-f54ln_f7a19182-b617-47e3-bc7f-db372257b984/barbican-worker/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.288319 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7c7c4b9c77-f54ln_f7a19182-b617-47e3-bc7f-db372257b984/barbican-worker-log/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.495751 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl_ecded0ed-1510-4ac8-80d0-dd7b06c8446e/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.548139 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6330b5b-9599-49b6-89fa-b890374a7830/ceilometer-central-agent/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.660484 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6330b5b-9599-49b6-89fa-b890374a7830/ceilometer-notification-agent/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.723278 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6330b5b-9599-49b6-89fa-b890374a7830/sg-core/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.738278 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6330b5b-9599-49b6-89fa-b890374a7830/proxy-httpd/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.925829 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d33ca8c1-d71d-4624-a362-f845fda14976/cinder-api/0.log" Dec 03 14:15:43 crc kubenswrapper[4690]: I1203 14:15:43.975599 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d33ca8c1-d71d-4624-a362-f845fda14976/cinder-api-log/0.log" Dec 03 14:15:44 crc kubenswrapper[4690]: I1203 14:15:44.078336 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fa834771-9deb-40da-b5ae-4f8d60667d2b/cinder-scheduler/0.log" Dec 03 14:15:44 crc kubenswrapper[4690]: I1203 14:15:44.765323 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fa834771-9deb-40da-b5ae-4f8d60667d2b/probe/0.log" Dec 03 14:15:44 crc kubenswrapper[4690]: I1203 14:15:44.779061 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-p55xh_69f0b3dc-dc18-4a71-ad0c-a5023db9ca96/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:44 crc kubenswrapper[4690]: I1203 14:15:44.935241 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-77rvl_ede30151-25e0-47e4-98b4-bf7ca4d55255/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.008747 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-wn8bx_fa48f557-98dd-49ca-a4a6-89b196b77b15/init/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.212713 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-wn8bx_fa48f557-98dd-49ca-a4a6-89b196b77b15/init/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.256672 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-wn8bx_fa48f557-98dd-49ca-a4a6-89b196b77b15/dnsmasq-dns/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.308192 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pksbh_2f30c127-ad2e-4154-a344-61a21b57b3d0/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.531396 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ada5ac2f-e8b5-4437-b639-183fe2bf3c09/glance-log/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.580813 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ada5ac2f-e8b5-4437-b639-183fe2bf3c09/glance-httpd/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.734830 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8f2638db-cd46-439c-bf63-4cd912784601/glance-log/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.741489 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8f2638db-cd46-439c-bf63-4cd912784601/glance-httpd/0.log" Dec 03 14:15:45 crc kubenswrapper[4690]: I1203 14:15:45.783341 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr_b529c361-6189-47c6-9c62-ea7fc0e8e30f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:46 crc kubenswrapper[4690]: I1203 14:15:46.013482 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kglnb_de573b60-ca9e-483c-8b4a-6d62fc6f56eb/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:46 crc kubenswrapper[4690]: I1203 14:15:46.258304 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412841-vn5nx_99ea4ce8-3695-4f3d-97fd-99d0e0b87029/keystone-cron/0.log" Dec 03 14:15:46 crc kubenswrapper[4690]: I1203 14:15:46.478441 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-788df6b698-7gkrp_cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4/keystone-api/0.log" Dec 03 14:15:46 crc kubenswrapper[4690]: I1203 14:15:46.913529 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9869e3c2-67c6-47be-8f00-45999ffcd3e7/kube-state-metrics/0.log" Dec 03 14:15:47 crc kubenswrapper[4690]: I1203 14:15:47.058350 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-tgvwl_f66f8700-77ac-4bed-b21b-8866f0edc58b/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:47 crc kubenswrapper[4690]: I1203 14:15:47.068105 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq_8ce19539-f48f-446b-8a79-d528d58a9e8d/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:47 crc kubenswrapper[4690]: I1203 14:15:47.472712 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d87869f9-cmzfc_7d486acd-1716-4926-be13-3e67b3106c02/neutron-httpd/0.log" Dec 03 14:15:47 crc kubenswrapper[4690]: I1203 14:15:47.544568 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz_b1d5dc5c-d0a9-4194-88df-b48e39cad7d8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:47 crc kubenswrapper[4690]: I1203 14:15:47.580925 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d87869f9-cmzfc_7d486acd-1716-4926-be13-3e67b3106c02/neutron-api/0.log" Dec 03 14:15:48 crc kubenswrapper[4690]: I1203 14:15:48.045663 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278/nova-api-log/0.log" Dec 03 14:15:48 crc kubenswrapper[4690]: I1203 14:15:48.123959 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a730fe99-3893-4697-b946-f58470a93407/nova-cell0-conductor-conductor/0.log" Dec 03 14:15:48 crc kubenswrapper[4690]: I1203 14:15:48.182144 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278/nova-api-api/0.log" Dec 03 14:15:48 crc kubenswrapper[4690]: I1203 14:15:48.389691 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_3141a41c-66ea-41fe-bc21-dd115b3cb98e/nova-cell1-conductor-conductor/0.log" Dec 03 14:15:48 crc kubenswrapper[4690]: I1203 14:15:48.485579 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d9db005b-de33-4de4-b7fb-c60783700345/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 14:15:48 crc kubenswrapper[4690]: I1203 14:15:48.675994 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hgcm8_6b06fe71-66c9-48a8-9e96-b9bdb23daf92/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:48 crc kubenswrapper[4690]: I1203 14:15:48.841923 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3f0b2df1-b5ed-498c-b6dd-784a32f97b3f/nova-metadata-log/0.log" Dec 03 14:15:49 crc kubenswrapper[4690]: I1203 14:15:49.076959 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_cc0fda9d-61ad-4a67-8c22-1da2e41949b0/nova-scheduler-scheduler/0.log" Dec 03 14:15:49 crc kubenswrapper[4690]: I1203 14:15:49.141418 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c157eb34-b68f-4510-aafd-0c5e73435efe/mysql-bootstrap/0.log" Dec 03 14:15:49 crc kubenswrapper[4690]: I1203 14:15:49.440593 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c157eb34-b68f-4510-aafd-0c5e73435efe/galera/0.log" Dec 03 14:15:49 crc kubenswrapper[4690]: I1203 14:15:49.446992 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c157eb34-b68f-4510-aafd-0c5e73435efe/mysql-bootstrap/0.log" Dec 03 14:15:49 crc kubenswrapper[4690]: I1203 14:15:49.731573 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5d47d61-0012-4eb8-ba88-da731d54ad45/mysql-bootstrap/0.log" Dec 03 14:15:49 crc kubenswrapper[4690]: I1203 14:15:49.862024 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5d47d61-0012-4eb8-ba88-da731d54ad45/mysql-bootstrap/0.log" Dec 03 14:15:49 crc kubenswrapper[4690]: I1203 14:15:49.954488 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5d47d61-0012-4eb8-ba88-da731d54ad45/galera/0.log" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.058676 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f47b845f-82f3-44b9-966d-eb3076f05362/openstackclient/0.log" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.230276 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-wjlxx_338580e8-66c9-4169-93f5-1e7f530de31c/openstack-network-exporter/0.log" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.282797 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3f0b2df1-b5ed-498c-b6dd-784a32f97b3f/nova-metadata-metadata/0.log" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.316182 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:15:50 crc kubenswrapper[4690]: E1203 14:15:50.316417 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.454769 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-mf8xw_64710eed-cdda-4f62-a007-766dd92f0b29/ovn-controller/0.log" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.468018 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-znnxj_cf553f8a-b61b-469b-9c0c-d602d73ef4c2/ovsdb-server-init/0.log" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.753577 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-znnxj_cf553f8a-b61b-469b-9c0c-d602d73ef4c2/ovsdb-server-init/0.log" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.838493 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-znnxj_cf553f8a-b61b-469b-9c0c-d602d73ef4c2/ovs-vswitchd/0.log" Dec 03 14:15:50 crc kubenswrapper[4690]: I1203 14:15:50.876346 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-znnxj_cf553f8a-b61b-469b-9c0c-d602d73ef4c2/ovsdb-server/0.log" Dec 03 14:15:51 crc kubenswrapper[4690]: I1203 14:15:51.085351 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a172f1ad-a0fb-4f37-92e0-2c7267d11fa8/openstack-network-exporter/0.log" Dec 03 14:15:51 crc kubenswrapper[4690]: I1203 14:15:51.150271 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-6pqsk_2ba8f301-845c-4425-9a12-72babfccd551/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:51 crc kubenswrapper[4690]: I1203 14:15:51.210245 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a172f1ad-a0fb-4f37-92e0-2c7267d11fa8/ovn-northd/0.log" Dec 03 14:15:51 crc kubenswrapper[4690]: I1203 14:15:51.396638 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c44ebdaa-211b-4ac6-90c6-5236eaee6295/openstack-network-exporter/0.log" Dec 03 14:15:51 crc kubenswrapper[4690]: I1203 14:15:51.463120 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c44ebdaa-211b-4ac6-90c6-5236eaee6295/ovsdbserver-nb/0.log" Dec 03 14:15:51 crc kubenswrapper[4690]: I1203 14:15:51.633371 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d19eec48-0864-43d8-8071-1aef21a26286/openstack-network-exporter/0.log" Dec 03 14:15:51 crc kubenswrapper[4690]: I1203 14:15:51.716276 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d19eec48-0864-43d8-8071-1aef21a26286/ovsdbserver-sb/0.log" Dec 03 14:15:51 crc kubenswrapper[4690]: I1203 14:15:51.819485 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f775c89b8-4tcrc_f9a732dd-555c-4e51-a334-80dab3104192/placement-api/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.022508 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f775c89b8-4tcrc_f9a732dd-555c-4e51-a334-80dab3104192/placement-log/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.049640 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8daef7dd-5b47-4f5c-8dc8-0da54d682e1d/setup-container/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.301547 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8daef7dd-5b47-4f5c-8dc8-0da54d682e1d/setup-container/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.386137 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8daef7dd-5b47-4f5c-8dc8-0da54d682e1d/rabbitmq/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.443673 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8/setup-container/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.628017 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8/setup-container/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.671979 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8/rabbitmq/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.725889 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt_0d10ad8c-5b1c-45c2-890a-eb42c648d437/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.919481 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-f924x_d4625caf-d841-4682-ae69-e7d2d90ef220/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:52 crc kubenswrapper[4690]: I1203 14:15:52.982368 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm_154bbd71-1585-49b0-84ab-3d11d49c99e6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:53 crc kubenswrapper[4690]: I1203 14:15:53.253262 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-2qlb4_80400476-b39e-4e4f-b57f-8813951d0f1a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:53 crc kubenswrapper[4690]: I1203 14:15:53.403709 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-gvv8v_91d255ec-88b0-416f-a990-92d930335b46/ssh-known-hosts-edpm-deployment/0.log" Dec 03 14:15:53 crc kubenswrapper[4690]: I1203 14:15:53.562292 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696775c5d7-q6gwf_9975bb19-d67e-4a25-a69f-8c1e37653961/proxy-server/0.log" Dec 03 14:15:53 crc kubenswrapper[4690]: I1203 14:15:53.707809 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bl55c_262bf571-0151-4c11-a05c-1270656144f9/swift-ring-rebalance/0.log" Dec 03 14:15:53 crc kubenswrapper[4690]: I1203 14:15:53.761250 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696775c5d7-q6gwf_9975bb19-d67e-4a25-a69f-8c1e37653961/proxy-httpd/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.136549 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/account-reaper/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.160379 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/account-auditor/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.231084 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/account-replicator/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.316971 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/account-server/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.374302 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/container-auditor/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.414850 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/container-replicator/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.494794 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/container-server/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.623107 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/container-updater/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.661838 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-auditor/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.739478 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-expirer/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.775920 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-replicator/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.853693 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-server/0.log" Dec 03 14:15:54 crc kubenswrapper[4690]: I1203 14:15:54.960531 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-updater/0.log" Dec 03 14:15:55 crc kubenswrapper[4690]: I1203 14:15:55.018573 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/rsync/0.log" Dec 03 14:15:55 crc kubenswrapper[4690]: I1203 14:15:55.024457 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_187a8435-f796-4435-a0b0-211daf924823/memcached/0.log" Dec 03 14:15:55 crc kubenswrapper[4690]: I1203 14:15:55.031796 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/swift-recon-cron/0.log" Dec 03 14:15:55 crc kubenswrapper[4690]: I1203 14:15:55.198674 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n_c21f8ec4-b27d-4a89-bd74-e03efa0afe48/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:15:55 crc kubenswrapper[4690]: I1203 14:15:55.254806 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_46b67a35-a83b-48dd-a6ea-4464ff4ee3d5/tempest-tests-tempest-tests-runner/0.log" Dec 03 14:15:55 crc kubenswrapper[4690]: I1203 14:15:55.393686 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_7b33ef02-9b68-41ad-90ed-1adfe0cd8312/test-operator-logs-container/0.log" Dec 03 14:15:55 crc kubenswrapper[4690]: I1203 14:15:55.467442 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr_96b45681-d5bf-474b-ac2c-2f8df92494d9/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:16:01 crc kubenswrapper[4690]: I1203 14:16:01.316977 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:16:01 crc kubenswrapper[4690]: E1203 14:16:01.319523 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:16:13 crc kubenswrapper[4690]: I1203 14:16:13.316111 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:16:13 crc kubenswrapper[4690]: E1203 14:16:13.317304 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:16:20 crc kubenswrapper[4690]: I1203 14:16:20.897120 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/util/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.138004 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/util/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.166947 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/pull/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.182534 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/pull/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.364122 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/pull/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.371070 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/util/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.400970 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/extract/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.597917 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-hx22v_1239bbbc-6613-4e1e-867b-fd62dee3c32b/kube-rbac-proxy/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.668408 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pnhhq_bc776d95-7342-47e3-b2ad-abc4a3f7aa23/kube-rbac-proxy/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.795561 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-hx22v_1239bbbc-6613-4e1e-867b-fd62dee3c32b/manager/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.882181 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pnhhq_bc776d95-7342-47e3-b2ad-abc4a3f7aa23/manager/0.log" Dec 03 14:16:21 crc kubenswrapper[4690]: I1203 14:16:21.936142 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-hwxjb_02c9e233-e41c-4fd2-9165-0ad5413fdaa2/kube-rbac-proxy/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.013505 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-hwxjb_02c9e233-e41c-4fd2-9165-0ad5413fdaa2/manager/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.096543 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-jmz4q_3b3afb5d-aef4-4c05-876f-52fa03532fe1/kube-rbac-proxy/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.237990 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-jmz4q_3b3afb5d-aef4-4c05-876f-52fa03532fe1/manager/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.370609 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-k6ht5_9da3821b-8edf-46c2-ac1c-5cabd4124968/kube-rbac-proxy/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.395096 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-k6ht5_9da3821b-8edf-46c2-ac1c-5cabd4124968/manager/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.564588 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gtcwg_4a193df2-0e41-48f0-a997-6f4b4e65bfe6/kube-rbac-proxy/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.634417 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gtcwg_4a193df2-0e41-48f0-a997-6f4b4e65bfe6/manager/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.689812 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-rrqhz_729e315d-2bf3-4568-941e-98ef7ad281eb/kube-rbac-proxy/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.900171 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-kshv8_5f4f6cc8-0d38-4c2a-b4f5-10a870c48479/kube-rbac-proxy/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.949124 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-kshv8_5f4f6cc8-0d38-4c2a-b4f5-10a870c48479/manager/0.log" Dec 03 14:16:22 crc kubenswrapper[4690]: I1203 14:16:22.961203 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-rrqhz_729e315d-2bf3-4568-941e-98ef7ad281eb/manager/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.125755 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-49f9d_0e42df6f-0109-4a0b-afeb-630bff5b8c35/kube-rbac-proxy/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.206744 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-49f9d_0e42df6f-0109-4a0b-afeb-630bff5b8c35/manager/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.350274 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-jk4h5_c731a6c0-1c65-45e1-893d-f620e83b5363/kube-rbac-proxy/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.384726 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-jk4h5_c731a6c0-1c65-45e1-893d-f620e83b5363/manager/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.513734 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-6g8xn_abb876f0-132b-4224-a3a7-074beba3367b/kube-rbac-proxy/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.590746 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-6g8xn_abb876f0-132b-4224-a3a7-074beba3367b/manager/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.639646 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-p559w_87cb4391-ed76-4f7a-a4b5-481d409bba7d/kube-rbac-proxy/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.784253 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-p559w_87cb4391-ed76-4f7a-a4b5-481d409bba7d/manager/0.log" Dec 03 14:16:23 crc kubenswrapper[4690]: I1203 14:16:23.903155 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5fll5_942df52e-3f08-47ef-bbeb-fe1191ced3e0/kube-rbac-proxy/0.log" Dec 03 14:16:24 crc kubenswrapper[4690]: I1203 14:16:24.015675 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5fll5_942df52e-3f08-47ef-bbeb-fe1191ced3e0/manager/0.log" Dec 03 14:16:24 crc kubenswrapper[4690]: I1203 14:16:24.072430 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-6q5ss_9890f67c-68ed-475b-9089-262d1fc270b4/kube-rbac-proxy/0.log" Dec 03 14:16:24 crc kubenswrapper[4690]: I1203 14:16:24.159403 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-6q5ss_9890f67c-68ed-475b-9089-262d1fc270b4/manager/0.log" Dec 03 14:16:24 crc kubenswrapper[4690]: I1203 14:16:24.306091 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd_f8b73cbf-3a07-46fa-98a4-db492ee1b711/kube-rbac-proxy/0.log" Dec 03 14:16:24 crc kubenswrapper[4690]: I1203 14:16:24.395186 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd_f8b73cbf-3a07-46fa-98a4-db492ee1b711/manager/0.log" Dec 03 14:16:24 crc kubenswrapper[4690]: I1203 14:16:24.762029 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-95zhf_ff4c17ee-a756-4743-aeb8-aae8fe520712/registry-server/0.log" Dec 03 14:16:24 crc kubenswrapper[4690]: I1203 14:16:24.900989 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-58bc846945-vrqwj_b3182947-b66b-4777-8410-b6ffc37aa5d0/operator/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.068378 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lmq2n_7ef3f1c5-17d7-42f0-949b-1dbab5568051/manager/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.068854 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lmq2n_7ef3f1c5-17d7-42f0-949b-1dbab5568051/kube-rbac-proxy/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.220012 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7xsmn_55e64346-52e8-49b2-81de-057a98be5150/kube-rbac-proxy/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.322289 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:16:25 crc kubenswrapper[4690]: E1203 14:16:25.323054 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.323760 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7xsmn_55e64346-52e8-49b2-81de-057a98be5150/manager/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.426354 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-fjthl_eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58/operator/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.600783 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6xs6t_00aeb5e7-0973-4852-9fcf-6a51c3ef9720/kube-rbac-proxy/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.662296 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6xs6t_00aeb5e7-0973-4852-9fcf-6a51c3ef9720/manager/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.784302 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d8d6bf864-kzk67_7df25165-6b23-4dda-8105-05a933bc3ac8/kube-rbac-proxy/0.log" Dec 03 14:16:25 crc kubenswrapper[4690]: I1203 14:16:25.945082 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d8d6bf864-kzk67_7df25165-6b23-4dda-8105-05a933bc3ac8/manager/0.log" Dec 03 14:16:26 crc kubenswrapper[4690]: I1203 14:16:26.000193 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-x7xbr_0e2ce38d-747d-408f-8768-48a7310c0f9c/kube-rbac-proxy/0.log" Dec 03 14:16:26 crc kubenswrapper[4690]: I1203 14:16:26.000247 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7874b9874-9jdm5_953dae05-293d-488a-96d3-af4eda42ad47/manager/0.log" Dec 03 14:16:26 crc kubenswrapper[4690]: I1203 14:16:26.027330 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-x7xbr_0e2ce38d-747d-408f-8768-48a7310c0f9c/manager/0.log" Dec 03 14:16:26 crc kubenswrapper[4690]: I1203 14:16:26.144294 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-skzwp_172f574f-c4bf-450a-8da9-fb4b879275f9/kube-rbac-proxy/0.log" Dec 03 14:16:26 crc kubenswrapper[4690]: I1203 14:16:26.173577 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-skzwp_172f574f-c4bf-450a-8da9-fb4b879275f9/manager/0.log" Dec 03 14:16:39 crc kubenswrapper[4690]: I1203 14:16:39.315771 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:16:39 crc kubenswrapper[4690]: E1203 14:16:39.317739 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:16:46 crc kubenswrapper[4690]: I1203 14:16:46.600838 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jvcdk_2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3/control-plane-machine-set-operator/0.log" Dec 03 14:16:46 crc kubenswrapper[4690]: I1203 14:16:46.784840 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jldld_fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26/machine-api-operator/0.log" Dec 03 14:16:46 crc kubenswrapper[4690]: I1203 14:16:46.824767 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jldld_fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26/kube-rbac-proxy/0.log" Dec 03 14:16:52 crc kubenswrapper[4690]: I1203 14:16:52.315848 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:16:52 crc kubenswrapper[4690]: E1203 14:16:52.318932 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:16:59 crc kubenswrapper[4690]: I1203 14:16:59.470465 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2gp2s_9c22c662-9a32-4721-9958-12651fa443a7/cert-manager-controller/0.log" Dec 03 14:16:59 crc kubenswrapper[4690]: I1203 14:16:59.678394 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-2p8k2_6daa0e47-6f3f-4ac8-b1a8-596401c33ef2/cert-manager-cainjector/0.log" Dec 03 14:16:59 crc kubenswrapper[4690]: I1203 14:16:59.702749 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4jfnl_f47a381d-4adf-4543-b4f3-793cdc3db649/cert-manager-webhook/0.log" Dec 03 14:17:07 crc kubenswrapper[4690]: I1203 14:17:07.315377 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:17:07 crc kubenswrapper[4690]: E1203 14:17:07.316286 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.499710 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tp4lg"] Dec 03 14:17:10 crc kubenswrapper[4690]: E1203 14:17:10.501402 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5521c45e-10ae-465e-8e25-628e67182be1" containerName="container-00" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.501424 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="5521c45e-10ae-465e-8e25-628e67182be1" containerName="container-00" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.501745 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="5521c45e-10ae-465e-8e25-628e67182be1" containerName="container-00" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.503717 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.524394 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tp4lg"] Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.613543 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-utilities\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.613981 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-catalog-content\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.614038 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qmn5\" (UniqueName: \"kubernetes.io/projected/a9a8c761-8142-4ad4-ba94-ce9480d07170-kube-api-access-4qmn5\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.716112 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-utilities\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.716190 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-catalog-content\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.716257 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qmn5\" (UniqueName: \"kubernetes.io/projected/a9a8c761-8142-4ad4-ba94-ce9480d07170-kube-api-access-4qmn5\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.716936 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-utilities\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.716936 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-catalog-content\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.737906 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qmn5\" (UniqueName: \"kubernetes.io/projected/a9a8c761-8142-4ad4-ba94-ce9480d07170-kube-api-access-4qmn5\") pod \"community-operators-tp4lg\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:10 crc kubenswrapper[4690]: I1203 14:17:10.835725 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:11 crc kubenswrapper[4690]: W1203 14:17:11.345965 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a8c761_8142_4ad4_ba94_ce9480d07170.slice/crio-3ef2635c3c35c66b2edf1dd1e896146a86d9fe2297549dd36434fec3a09578dc WatchSource:0}: Error finding container 3ef2635c3c35c66b2edf1dd1e896146a86d9fe2297549dd36434fec3a09578dc: Status 404 returned error can't find the container with id 3ef2635c3c35c66b2edf1dd1e896146a86d9fe2297549dd36434fec3a09578dc Dec 03 14:17:11 crc kubenswrapper[4690]: I1203 14:17:11.346801 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tp4lg"] Dec 03 14:17:12 crc kubenswrapper[4690]: I1203 14:17:12.090140 4690 generic.go:334] "Generic (PLEG): container finished" podID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerID="0494ec257ac16fb2752d8f4c17aad78afc55a70199a2fb0b90f422b53ef2abfa" exitCode=0 Dec 03 14:17:12 crc kubenswrapper[4690]: I1203 14:17:12.090201 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp4lg" event={"ID":"a9a8c761-8142-4ad4-ba94-ce9480d07170","Type":"ContainerDied","Data":"0494ec257ac16fb2752d8f4c17aad78afc55a70199a2fb0b90f422b53ef2abfa"} Dec 03 14:17:12 crc kubenswrapper[4690]: I1203 14:17:12.090542 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp4lg" event={"ID":"a9a8c761-8142-4ad4-ba94-ce9480d07170","Type":"ContainerStarted","Data":"3ef2635c3c35c66b2edf1dd1e896146a86d9fe2297549dd36434fec3a09578dc"} Dec 03 14:17:13 crc kubenswrapper[4690]: I1203 14:17:13.598588 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-4xg2d_bb3da90d-da37-49ad-ad46-230759602e46/nmstate-console-plugin/0.log" Dec 03 14:17:13 crc kubenswrapper[4690]: I1203 14:17:13.744983 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-v429n_89d8054c-b35f-42fd-960d-16e1b39ad983/nmstate-handler/0.log" Dec 03 14:17:13 crc kubenswrapper[4690]: I1203 14:17:13.830399 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-hjrl4_33310d90-a18c-44db-a7f3-37f2086de63a/nmstate-metrics/0.log" Dec 03 14:17:13 crc kubenswrapper[4690]: I1203 14:17:13.840427 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-hjrl4_33310d90-a18c-44db-a7f3-37f2086de63a/kube-rbac-proxy/0.log" Dec 03 14:17:13 crc kubenswrapper[4690]: I1203 14:17:13.952124 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-p69d4_0aacb971-67b1-4e6a-a89c-7aa8b404f046/nmstate-operator/0.log" Dec 03 14:17:14 crc kubenswrapper[4690]: I1203 14:17:14.078011 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-mw7tn_6020072e-2f65-4d13-85be-d300b4ec7f6d/nmstate-webhook/0.log" Dec 03 14:17:14 crc kubenswrapper[4690]: I1203 14:17:14.115156 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp4lg" event={"ID":"a9a8c761-8142-4ad4-ba94-ce9480d07170","Type":"ContainerStarted","Data":"a5354a0f93536977e3280c679a0bc6b268f3e63572c899a33cc2adfcd260441c"} Dec 03 14:17:15 crc kubenswrapper[4690]: I1203 14:17:15.129983 4690 generic.go:334] "Generic (PLEG): container finished" podID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerID="a5354a0f93536977e3280c679a0bc6b268f3e63572c899a33cc2adfcd260441c" exitCode=0 Dec 03 14:17:15 crc kubenswrapper[4690]: I1203 14:17:15.130104 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp4lg" event={"ID":"a9a8c761-8142-4ad4-ba94-ce9480d07170","Type":"ContainerDied","Data":"a5354a0f93536977e3280c679a0bc6b268f3e63572c899a33cc2adfcd260441c"} Dec 03 14:17:17 crc kubenswrapper[4690]: I1203 14:17:17.157657 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp4lg" event={"ID":"a9a8c761-8142-4ad4-ba94-ce9480d07170","Type":"ContainerStarted","Data":"915c51a9a83013c86ffe8c19593113afb05974b2290c9f30982e8ff100f59d2f"} Dec 03 14:17:17 crc kubenswrapper[4690]: I1203 14:17:17.190138 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tp4lg" podStartSLOduration=3.655557446 podStartE2EDuration="7.190098097s" podCreationTimestamp="2025-12-03 14:17:10 +0000 UTC" firstStartedPulling="2025-12-03 14:17:12.092084364 +0000 UTC m=+4078.073004807" lastFinishedPulling="2025-12-03 14:17:15.626625025 +0000 UTC m=+4081.607545458" observedRunningTime="2025-12-03 14:17:17.180949196 +0000 UTC m=+4083.161869649" watchObservedRunningTime="2025-12-03 14:17:17.190098097 +0000 UTC m=+4083.171018530" Dec 03 14:17:18 crc kubenswrapper[4690]: I1203 14:17:18.314945 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:17:18 crc kubenswrapper[4690]: E1203 14:17:18.315624 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:17:20 crc kubenswrapper[4690]: I1203 14:17:20.836886 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:20 crc kubenswrapper[4690]: I1203 14:17:20.837305 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:20 crc kubenswrapper[4690]: I1203 14:17:20.895158 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:21 crc kubenswrapper[4690]: I1203 14:17:21.246216 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:21 crc kubenswrapper[4690]: I1203 14:17:21.306277 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tp4lg"] Dec 03 14:17:23 crc kubenswrapper[4690]: I1203 14:17:23.213522 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tp4lg" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerName="registry-server" containerID="cri-o://915c51a9a83013c86ffe8c19593113afb05974b2290c9f30982e8ff100f59d2f" gracePeriod=2 Dec 03 14:17:23 crc kubenswrapper[4690]: E1203 14:17:23.473898 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a8c761_8142_4ad4_ba94_ce9480d07170.slice/crio-915c51a9a83013c86ffe8c19593113afb05974b2290c9f30982e8ff100f59d2f.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.235535 4690 generic.go:334] "Generic (PLEG): container finished" podID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerID="915c51a9a83013c86ffe8c19593113afb05974b2290c9f30982e8ff100f59d2f" exitCode=0 Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.235697 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp4lg" event={"ID":"a9a8c761-8142-4ad4-ba94-ce9480d07170","Type":"ContainerDied","Data":"915c51a9a83013c86ffe8c19593113afb05974b2290c9f30982e8ff100f59d2f"} Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.357276 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.540647 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qmn5\" (UniqueName: \"kubernetes.io/projected/a9a8c761-8142-4ad4-ba94-ce9480d07170-kube-api-access-4qmn5\") pod \"a9a8c761-8142-4ad4-ba94-ce9480d07170\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.540986 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-utilities\") pod \"a9a8c761-8142-4ad4-ba94-ce9480d07170\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.541054 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-catalog-content\") pod \"a9a8c761-8142-4ad4-ba94-ce9480d07170\" (UID: \"a9a8c761-8142-4ad4-ba94-ce9480d07170\") " Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.541733 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-utilities" (OuterVolumeSpecName: "utilities") pod "a9a8c761-8142-4ad4-ba94-ce9480d07170" (UID: "a9a8c761-8142-4ad4-ba94-ce9480d07170"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.547231 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9a8c761-8142-4ad4-ba94-ce9480d07170-kube-api-access-4qmn5" (OuterVolumeSpecName: "kube-api-access-4qmn5") pod "a9a8c761-8142-4ad4-ba94-ce9480d07170" (UID: "a9a8c761-8142-4ad4-ba94-ce9480d07170"). InnerVolumeSpecName "kube-api-access-4qmn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.609356 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9a8c761-8142-4ad4-ba94-ce9480d07170" (UID: "a9a8c761-8142-4ad4-ba94-ce9480d07170"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.643616 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.643655 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a8c761-8142-4ad4-ba94-ce9480d07170-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:17:24 crc kubenswrapper[4690]: I1203 14:17:24.643672 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qmn5\" (UniqueName: \"kubernetes.io/projected/a9a8c761-8142-4ad4-ba94-ce9480d07170-kube-api-access-4qmn5\") on node \"crc\" DevicePath \"\"" Dec 03 14:17:25 crc kubenswrapper[4690]: I1203 14:17:25.249104 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp4lg" event={"ID":"a9a8c761-8142-4ad4-ba94-ce9480d07170","Type":"ContainerDied","Data":"3ef2635c3c35c66b2edf1dd1e896146a86d9fe2297549dd36434fec3a09578dc"} Dec 03 14:17:25 crc kubenswrapper[4690]: I1203 14:17:25.249166 4690 scope.go:117] "RemoveContainer" containerID="915c51a9a83013c86ffe8c19593113afb05974b2290c9f30982e8ff100f59d2f" Dec 03 14:17:25 crc kubenswrapper[4690]: I1203 14:17:25.249192 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp4lg" Dec 03 14:17:25 crc kubenswrapper[4690]: I1203 14:17:25.289059 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tp4lg"] Dec 03 14:17:25 crc kubenswrapper[4690]: I1203 14:17:25.296206 4690 scope.go:117] "RemoveContainer" containerID="a5354a0f93536977e3280c679a0bc6b268f3e63572c899a33cc2adfcd260441c" Dec 03 14:17:25 crc kubenswrapper[4690]: I1203 14:17:25.304202 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tp4lg"] Dec 03 14:17:25 crc kubenswrapper[4690]: I1203 14:17:25.327857 4690 scope.go:117] "RemoveContainer" containerID="0494ec257ac16fb2752d8f4c17aad78afc55a70199a2fb0b90f422b53ef2abfa" Dec 03 14:17:26 crc kubenswrapper[4690]: I1203 14:17:26.328113 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" path="/var/lib/kubelet/pods/a9a8c761-8142-4ad4-ba94-ce9480d07170/volumes" Dec 03 14:17:28 crc kubenswrapper[4690]: I1203 14:17:28.110106 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-bb5b4b448-ql656_72a896a4-41ed-4652-bfa5-358369eddfa2/kube-rbac-proxy/0.log" Dec 03 14:17:28 crc kubenswrapper[4690]: I1203 14:17:28.187221 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-bb5b4b448-ql656_72a896a4-41ed-4652-bfa5-358369eddfa2/manager/0.log" Dec 03 14:17:29 crc kubenswrapper[4690]: I1203 14:17:29.315418 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:17:29 crc kubenswrapper[4690]: E1203 14:17:29.315811 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:17:43 crc kubenswrapper[4690]: I1203 14:17:43.315312 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:17:43 crc kubenswrapper[4690]: E1203 14:17:43.316407 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:17:43 crc kubenswrapper[4690]: I1203 14:17:43.378803 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-lqf4z_fb39061a-2c4c-4c29-95b6-6f070367df86/cluster-logging-operator/0.log" Dec 03 14:17:43 crc kubenswrapper[4690]: I1203 14:17:43.564375 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-g89tt_ed87f330-d809-4183-8cc7-61caf74208a8/collector/0.log" Dec 03 14:17:43 crc kubenswrapper[4690]: I1203 14:17:43.571647 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_6289c341-366c-4b2e-a4ed-2d0eb43d0da4/loki-compactor/0.log" Dec 03 14:17:43 crc kubenswrapper[4690]: I1203 14:17:43.752614 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-qx2hh_d4eb66c8-ca39-4ba9-8717-a488817c5304/loki-distributor/0.log" Dec 03 14:17:43 crc kubenswrapper[4690]: I1203 14:17:43.831545 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-79d97f9547-8tgpn_4b3976e3-672e-4117-a9c6-9a7f34d42be7/opa/0.log" Dec 03 14:17:43 crc kubenswrapper[4690]: I1203 14:17:43.837521 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-79d97f9547-8tgpn_4b3976e3-672e-4117-a9c6-9a7f34d42be7/gateway/0.log" Dec 03 14:17:44 crc kubenswrapper[4690]: I1203 14:17:44.036025 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-79d97f9547-sps84_f1541c44-fb2e-4414-88f0-21e464b629f0/gateway/0.log" Dec 03 14:17:44 crc kubenswrapper[4690]: I1203 14:17:44.118461 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-79d97f9547-sps84_f1541c44-fb2e-4414-88f0-21e464b629f0/opa/0.log" Dec 03 14:17:44 crc kubenswrapper[4690]: I1203 14:17:44.186836 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_e56039ef-9919-4d3d-9b04-876bad0f5756/loki-index-gateway/0.log" Dec 03 14:17:44 crc kubenswrapper[4690]: I1203 14:17:44.352568 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86/loki-ingester/0.log" Dec 03 14:17:44 crc kubenswrapper[4690]: I1203 14:17:44.398727 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-ckhv9_36825bd7-9921-4f76-a26d-6dd05f00bc3e/loki-querier/0.log" Dec 03 14:17:44 crc kubenswrapper[4690]: I1203 14:17:44.551565 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-pslms_1b8cb794-f99d-41b0-bbdc-ed052be8f4a5/loki-query-frontend/0.log" Dec 03 14:17:57 crc kubenswrapper[4690]: I1203 14:17:57.315213 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:17:57 crc kubenswrapper[4690]: E1203 14:17:57.315992 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:17:59 crc kubenswrapper[4690]: I1203 14:17:59.750506 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t4xcd_e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1/kube-rbac-proxy/0.log" Dec 03 14:17:59 crc kubenswrapper[4690]: I1203 14:17:59.898550 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t4xcd_e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1/controller/0.log" Dec 03 14:17:59 crc kubenswrapper[4690]: I1203 14:17:59.937684 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-frr-files/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.156269 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-frr-files/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.178579 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-reloader/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.190071 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-reloader/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.190241 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-metrics/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.401924 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-metrics/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.416078 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-reloader/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.421477 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-frr-files/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.438237 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-metrics/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.653739 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-frr-files/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.662681 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/controller/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.700698 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-reloader/0.log" Dec 03 14:18:00 crc kubenswrapper[4690]: I1203 14:18:00.722920 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-metrics/0.log" Dec 03 14:18:01 crc kubenswrapper[4690]: I1203 14:18:01.565162 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/frr-metrics/0.log" Dec 03 14:18:01 crc kubenswrapper[4690]: I1203 14:18:01.595067 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/kube-rbac-proxy/0.log" Dec 03 14:18:01 crc kubenswrapper[4690]: I1203 14:18:01.619020 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/kube-rbac-proxy-frr/0.log" Dec 03 14:18:01 crc kubenswrapper[4690]: I1203 14:18:01.805593 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/reloader/0.log" Dec 03 14:18:01 crc kubenswrapper[4690]: I1203 14:18:01.867890 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-r8zcv_ece18280-b324-4c1e-b3d5-411e5c146869/frr-k8s-webhook-server/0.log" Dec 03 14:18:02 crc kubenswrapper[4690]: I1203 14:18:02.144378 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5994f497d6-5c9m6_6b62fccc-0740-4e1f-8efb-61bac3a9261e/manager/0.log" Dec 03 14:18:02 crc kubenswrapper[4690]: I1203 14:18:02.377507 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c954657ff-b5h2m_93d03eb3-9a1d-452e-8812-c5a871652ab3/webhook-server/0.log" Dec 03 14:18:02 crc kubenswrapper[4690]: I1203 14:18:02.458298 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7hmjp_e1b3ec2d-089c-409a-901e-4f466ec07855/kube-rbac-proxy/0.log" Dec 03 14:18:02 crc kubenswrapper[4690]: I1203 14:18:02.905168 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/frr/0.log" Dec 03 14:18:03 crc kubenswrapper[4690]: I1203 14:18:03.251046 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7hmjp_e1b3ec2d-089c-409a-901e-4f466ec07855/speaker/0.log" Dec 03 14:18:12 crc kubenswrapper[4690]: I1203 14:18:12.315094 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:18:12 crc kubenswrapper[4690]: E1203 14:18:12.316018 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:18:17 crc kubenswrapper[4690]: I1203 14:18:17.574298 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/util/0.log" Dec 03 14:18:17 crc kubenswrapper[4690]: I1203 14:18:17.810490 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/pull/0.log" Dec 03 14:18:17 crc kubenswrapper[4690]: I1203 14:18:17.813212 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/util/0.log" Dec 03 14:18:17 crc kubenswrapper[4690]: I1203 14:18:17.827206 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/pull/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.001827 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/pull/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.036751 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/util/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.054827 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/extract/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.225422 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/util/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.368840 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/util/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.372253 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/pull/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.373834 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/pull/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.605501 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/pull/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.631684 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/util/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.658158 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/extract/0.log" Dec 03 14:18:18 crc kubenswrapper[4690]: I1203 14:18:18.801142 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/util/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.023015 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/pull/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.047746 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/util/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.050181 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/pull/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.226969 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/pull/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.228235 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/util/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.285876 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/extract/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.451271 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/util/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.665882 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/util/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.667555 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/pull/0.log" Dec 03 14:18:19 crc kubenswrapper[4690]: I1203 14:18:19.676173 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/pull/0.log" Dec 03 14:18:20 crc kubenswrapper[4690]: I1203 14:18:20.053803 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/util/0.log" Dec 03 14:18:20 crc kubenswrapper[4690]: I1203 14:18:20.073989 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/pull/0.log" Dec 03 14:18:20 crc kubenswrapper[4690]: I1203 14:18:20.134253 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/extract/0.log" Dec 03 14:18:20 crc kubenswrapper[4690]: I1203 14:18:20.626310 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-utilities/0.log" Dec 03 14:18:20 crc kubenswrapper[4690]: I1203 14:18:20.825587 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-content/0.log" Dec 03 14:18:20 crc kubenswrapper[4690]: I1203 14:18:20.867084 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-utilities/0.log" Dec 03 14:18:20 crc kubenswrapper[4690]: I1203 14:18:20.887154 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-content/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.104724 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-utilities/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.106636 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-content/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.456959 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-utilities/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.632070 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-content/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.690137 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-content/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.707768 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-utilities/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.718478 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/registry-server/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.947656 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-content/0.log" Dec 03 14:18:21 crc kubenswrapper[4690]: I1203 14:18:21.953793 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-utilities/0.log" Dec 03 14:18:22 crc kubenswrapper[4690]: I1203 14:18:22.027074 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-t8hvp_2b3e95ae-d895-45bf-9abd-73cf60e5f139/marketplace-operator/0.log" Dec 03 14:18:22 crc kubenswrapper[4690]: I1203 14:18:22.222720 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-utilities/0.log" Dec 03 14:18:22 crc kubenswrapper[4690]: I1203 14:18:22.882455 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/registry-server/0.log" Dec 03 14:18:22 crc kubenswrapper[4690]: I1203 14:18:22.892603 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-utilities/0.log" Dec 03 14:18:22 crc kubenswrapper[4690]: I1203 14:18:22.924259 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-content/0.log" Dec 03 14:18:22 crc kubenswrapper[4690]: I1203 14:18:22.928962 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-content/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.128442 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-utilities/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.182683 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-content/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.270623 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-utilities/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.298592 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/registry-server/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.407416 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-utilities/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.411943 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-content/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.467857 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-content/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.630382 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-utilities/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.661941 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-content/0.log" Dec 03 14:18:23 crc kubenswrapper[4690]: I1203 14:18:23.795517 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/registry-server/0.log" Dec 03 14:18:24 crc kubenswrapper[4690]: I1203 14:18:24.321970 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:18:25 crc kubenswrapper[4690]: I1203 14:18:25.012996 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"c2c14307ea04a24716f0f1945ac9b63fea01b9e0b074627d7758c3e1e8b30cd4"} Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.052698 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l6j5n"] Dec 03 14:18:32 crc kubenswrapper[4690]: E1203 14:18:32.054045 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerName="registry-server" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.054069 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerName="registry-server" Dec 03 14:18:32 crc kubenswrapper[4690]: E1203 14:18:32.054116 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerName="extract-content" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.054125 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerName="extract-content" Dec 03 14:18:32 crc kubenswrapper[4690]: E1203 14:18:32.054150 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerName="extract-utilities" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.054158 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerName="extract-utilities" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.054587 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9a8c761-8142-4ad4-ba94-ce9480d07170" containerName="registry-server" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.058002 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.066742 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l6j5n"] Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.217083 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpfnl\" (UniqueName: \"kubernetes.io/projected/15007eba-166c-4cf9-b4df-0e2a1410cce4-kube-api-access-bpfnl\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.217314 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-catalog-content\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.217942 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-utilities\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.320849 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpfnl\" (UniqueName: \"kubernetes.io/projected/15007eba-166c-4cf9-b4df-0e2a1410cce4-kube-api-access-bpfnl\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.321021 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-catalog-content\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.321222 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-utilities\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.321667 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-catalog-content\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.321700 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-utilities\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.347269 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpfnl\" (UniqueName: \"kubernetes.io/projected/15007eba-166c-4cf9-b4df-0e2a1410cce4-kube-api-access-bpfnl\") pod \"certified-operators-l6j5n\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.386606 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:32 crc kubenswrapper[4690]: I1203 14:18:32.944246 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l6j5n"] Dec 03 14:18:33 crc kubenswrapper[4690]: I1203 14:18:33.097300 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6j5n" event={"ID":"15007eba-166c-4cf9-b4df-0e2a1410cce4","Type":"ContainerStarted","Data":"f49925e790640f7d91b7be05b923c137d4fbf42555308f250be24e031a91db3e"} Dec 03 14:18:34 crc kubenswrapper[4690]: I1203 14:18:34.109372 4690 generic.go:334] "Generic (PLEG): container finished" podID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerID="322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6" exitCode=0 Dec 03 14:18:34 crc kubenswrapper[4690]: I1203 14:18:34.109485 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6j5n" event={"ID":"15007eba-166c-4cf9-b4df-0e2a1410cce4","Type":"ContainerDied","Data":"322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6"} Dec 03 14:18:34 crc kubenswrapper[4690]: I1203 14:18:34.112187 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:18:36 crc kubenswrapper[4690]: I1203 14:18:36.135196 4690 generic.go:334] "Generic (PLEG): container finished" podID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerID="12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca" exitCode=0 Dec 03 14:18:36 crc kubenswrapper[4690]: I1203 14:18:36.135284 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6j5n" event={"ID":"15007eba-166c-4cf9-b4df-0e2a1410cce4","Type":"ContainerDied","Data":"12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca"} Dec 03 14:18:37 crc kubenswrapper[4690]: I1203 14:18:37.150281 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6j5n" event={"ID":"15007eba-166c-4cf9-b4df-0e2a1410cce4","Type":"ContainerStarted","Data":"df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574"} Dec 03 14:18:37 crc kubenswrapper[4690]: I1203 14:18:37.176777 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l6j5n" podStartSLOduration=2.51629464 podStartE2EDuration="5.176749896s" podCreationTimestamp="2025-12-03 14:18:32 +0000 UTC" firstStartedPulling="2025-12-03 14:18:34.11175419 +0000 UTC m=+4160.092674623" lastFinishedPulling="2025-12-03 14:18:36.772209446 +0000 UTC m=+4162.753129879" observedRunningTime="2025-12-03 14:18:37.167142893 +0000 UTC m=+4163.148063336" watchObservedRunningTime="2025-12-03 14:18:37.176749896 +0000 UTC m=+4163.157670329" Dec 03 14:18:42 crc kubenswrapper[4690]: I1203 14:18:42.388149 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:42 crc kubenswrapper[4690]: I1203 14:18:42.388674 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:42 crc kubenswrapper[4690]: I1203 14:18:42.459830 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:43 crc kubenswrapper[4690]: I1203 14:18:43.602893 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:43 crc kubenswrapper[4690]: I1203 14:18:43.652921 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l6j5n"] Dec 03 14:18:45 crc kubenswrapper[4690]: I1203 14:18:45.252411 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l6j5n" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerName="registry-server" containerID="cri-o://df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574" gracePeriod=2 Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.141546 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.211316 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-catalog-content\") pod \"15007eba-166c-4cf9-b4df-0e2a1410cce4\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.211502 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpfnl\" (UniqueName: \"kubernetes.io/projected/15007eba-166c-4cf9-b4df-0e2a1410cce4-kube-api-access-bpfnl\") pod \"15007eba-166c-4cf9-b4df-0e2a1410cce4\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.211591 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-utilities\") pod \"15007eba-166c-4cf9-b4df-0e2a1410cce4\" (UID: \"15007eba-166c-4cf9-b4df-0e2a1410cce4\") " Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.212554 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-utilities" (OuterVolumeSpecName: "utilities") pod "15007eba-166c-4cf9-b4df-0e2a1410cce4" (UID: "15007eba-166c-4cf9-b4df-0e2a1410cce4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.218462 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15007eba-166c-4cf9-b4df-0e2a1410cce4-kube-api-access-bpfnl" (OuterVolumeSpecName: "kube-api-access-bpfnl") pod "15007eba-166c-4cf9-b4df-0e2a1410cce4" (UID: "15007eba-166c-4cf9-b4df-0e2a1410cce4"). InnerVolumeSpecName "kube-api-access-bpfnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.267781 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15007eba-166c-4cf9-b4df-0e2a1410cce4" (UID: "15007eba-166c-4cf9-b4df-0e2a1410cce4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.275375 4690 generic.go:334] "Generic (PLEG): container finished" podID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerID="df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574" exitCode=0 Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.275413 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6j5n" event={"ID":"15007eba-166c-4cf9-b4df-0e2a1410cce4","Type":"ContainerDied","Data":"df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574"} Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.275431 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l6j5n" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.275438 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6j5n" event={"ID":"15007eba-166c-4cf9-b4df-0e2a1410cce4","Type":"ContainerDied","Data":"f49925e790640f7d91b7be05b923c137d4fbf42555308f250be24e031a91db3e"} Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.275453 4690 scope.go:117] "RemoveContainer" containerID="df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.306328 4690 scope.go:117] "RemoveContainer" containerID="12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.314912 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpfnl\" (UniqueName: \"kubernetes.io/projected/15007eba-166c-4cf9-b4df-0e2a1410cce4-kube-api-access-bpfnl\") on node \"crc\" DevicePath \"\"" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.314949 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.314962 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15007eba-166c-4cf9-b4df-0e2a1410cce4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.326037 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l6j5n"] Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.332952 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l6j5n"] Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.351669 4690 scope.go:117] "RemoveContainer" containerID="322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.396772 4690 scope.go:117] "RemoveContainer" containerID="df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574" Dec 03 14:18:47 crc kubenswrapper[4690]: E1203 14:18:47.397621 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574\": container with ID starting with df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574 not found: ID does not exist" containerID="df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.397665 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574"} err="failed to get container status \"df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574\": rpc error: code = NotFound desc = could not find container \"df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574\": container with ID starting with df755d04e6be2d5173b1ad4aa1b4fcb7ccd4d130d5cd89f15b9a12465d2da574 not found: ID does not exist" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.397692 4690 scope.go:117] "RemoveContainer" containerID="12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca" Dec 03 14:18:47 crc kubenswrapper[4690]: E1203 14:18:47.397982 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca\": container with ID starting with 12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca not found: ID does not exist" containerID="12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.398006 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca"} err="failed to get container status \"12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca\": rpc error: code = NotFound desc = could not find container \"12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca\": container with ID starting with 12dc842971e988a1a7112b540ab930dca33ca13814a73c8c1895843c52ca95ca not found: ID does not exist" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.398018 4690 scope.go:117] "RemoveContainer" containerID="322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6" Dec 03 14:18:47 crc kubenswrapper[4690]: E1203 14:18:47.398217 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6\": container with ID starting with 322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6 not found: ID does not exist" containerID="322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6" Dec 03 14:18:47 crc kubenswrapper[4690]: I1203 14:18:47.398233 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6"} err="failed to get container status \"322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6\": rpc error: code = NotFound desc = could not find container \"322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6\": container with ID starting with 322621d002f17a355f413cf4a239d80aa79a8873019c22a01fb9acd1f3f8ebf6 not found: ID does not exist" Dec 03 14:18:48 crc kubenswrapper[4690]: I1203 14:18:48.327585 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" path="/var/lib/kubelet/pods/15007eba-166c-4cf9-b4df-0e2a1410cce4/volumes" Dec 03 14:18:49 crc kubenswrapper[4690]: I1203 14:18:49.653760 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-bb5b4b448-ql656_72a896a4-41ed-4652-bfa5-358369eddfa2/kube-rbac-proxy/0.log" Dec 03 14:18:49 crc kubenswrapper[4690]: I1203 14:18:49.729165 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-bb5b4b448-ql656_72a896a4-41ed-4652-bfa5-358369eddfa2/manager/0.log" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.118576 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x849l"] Dec 03 14:20:38 crc kubenswrapper[4690]: E1203 14:20:38.119754 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerName="extract-content" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.119767 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerName="extract-content" Dec 03 14:20:38 crc kubenswrapper[4690]: E1203 14:20:38.119814 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerName="registry-server" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.119820 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerName="registry-server" Dec 03 14:20:38 crc kubenswrapper[4690]: E1203 14:20:38.119833 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerName="extract-utilities" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.119843 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerName="extract-utilities" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.120121 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="15007eba-166c-4cf9-b4df-0e2a1410cce4" containerName="registry-server" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.121784 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.135924 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x849l"] Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.274983 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-catalog-content\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.275087 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7fb9\" (UniqueName: \"kubernetes.io/projected/795c8cf8-c96f-4bab-86d1-5881b53d40f9-kube-api-access-n7fb9\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.275175 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-utilities\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.377393 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-utilities\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.377499 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-catalog-content\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.377595 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7fb9\" (UniqueName: \"kubernetes.io/projected/795c8cf8-c96f-4bab-86d1-5881b53d40f9-kube-api-access-n7fb9\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.378579 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-utilities\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.378836 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-catalog-content\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.401990 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7fb9\" (UniqueName: \"kubernetes.io/projected/795c8cf8-c96f-4bab-86d1-5881b53d40f9-kube-api-access-n7fb9\") pod \"redhat-operators-x849l\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.452492 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.511546 4690 generic.go:334] "Generic (PLEG): container finished" podID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerID="e343e8a0a0ab1358feb642c0c16e1dcc26de2fbbd7c0c36b563162cefec04d7b" exitCode=0 Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.511634 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57qfs/must-gather-lzcsm" event={"ID":"cf253854-f14d-4ea6-9f4b-6e988d5fb432","Type":"ContainerDied","Data":"e343e8a0a0ab1358feb642c0c16e1dcc26de2fbbd7c0c36b563162cefec04d7b"} Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.512808 4690 scope.go:117] "RemoveContainer" containerID="e343e8a0a0ab1358feb642c0c16e1dcc26de2fbbd7c0c36b563162cefec04d7b" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.973424 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-57qfs_must-gather-lzcsm_cf253854-f14d-4ea6-9f4b-6e988d5fb432/gather/0.log" Dec 03 14:20:38 crc kubenswrapper[4690]: I1203 14:20:38.991931 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x849l"] Dec 03 14:20:39 crc kubenswrapper[4690]: I1203 14:20:39.526354 4690 generic.go:334] "Generic (PLEG): container finished" podID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerID="928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340" exitCode=0 Dec 03 14:20:39 crc kubenswrapper[4690]: I1203 14:20:39.526537 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x849l" event={"ID":"795c8cf8-c96f-4bab-86d1-5881b53d40f9","Type":"ContainerDied","Data":"928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340"} Dec 03 14:20:39 crc kubenswrapper[4690]: I1203 14:20:39.526777 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x849l" event={"ID":"795c8cf8-c96f-4bab-86d1-5881b53d40f9","Type":"ContainerStarted","Data":"3efe696d091068fab500c3b1fdc687f0355fc8c91f2628126bf861cabd747b02"} Dec 03 14:20:40 crc kubenswrapper[4690]: I1203 14:20:40.562822 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x849l" event={"ID":"795c8cf8-c96f-4bab-86d1-5881b53d40f9","Type":"ContainerStarted","Data":"4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6"} Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.310897 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8hcl6"] Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.315217 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.330648 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hcl6"] Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.360439 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-catalog-content\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.360495 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-utilities\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.360592 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg6md\" (UniqueName: \"kubernetes.io/projected/9364595b-9b85-47dd-adb3-357db3855ea6-kube-api-access-cg6md\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.465617 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-catalog-content\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.465704 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-utilities\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.465820 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg6md\" (UniqueName: \"kubernetes.io/projected/9364595b-9b85-47dd-adb3-357db3855ea6-kube-api-access-cg6md\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.466366 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-catalog-content\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.466473 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-utilities\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.489812 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg6md\" (UniqueName: \"kubernetes.io/projected/9364595b-9b85-47dd-adb3-357db3855ea6-kube-api-access-cg6md\") pod \"redhat-marketplace-8hcl6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:41 crc kubenswrapper[4690]: I1203 14:20:41.644762 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:42 crc kubenswrapper[4690]: I1203 14:20:42.202976 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hcl6"] Dec 03 14:20:42 crc kubenswrapper[4690]: I1203 14:20:42.586836 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hcl6" event={"ID":"9364595b-9b85-47dd-adb3-357db3855ea6","Type":"ContainerStarted","Data":"ea7ed38b933522142abdac3a949798914983132ad6367c1d484e00c32a466aab"} Dec 03 14:20:43 crc kubenswrapper[4690]: I1203 14:20:43.602207 4690 generic.go:334] "Generic (PLEG): container finished" podID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerID="4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6" exitCode=0 Dec 03 14:20:43 crc kubenswrapper[4690]: I1203 14:20:43.602258 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x849l" event={"ID":"795c8cf8-c96f-4bab-86d1-5881b53d40f9","Type":"ContainerDied","Data":"4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6"} Dec 03 14:20:43 crc kubenswrapper[4690]: I1203 14:20:43.604606 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hcl6" event={"ID":"9364595b-9b85-47dd-adb3-357db3855ea6","Type":"ContainerStarted","Data":"5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d"} Dec 03 14:20:45 crc kubenswrapper[4690]: I1203 14:20:45.626131 4690 generic.go:334] "Generic (PLEG): container finished" podID="9364595b-9b85-47dd-adb3-357db3855ea6" containerID="5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d" exitCode=0 Dec 03 14:20:45 crc kubenswrapper[4690]: I1203 14:20:45.626228 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hcl6" event={"ID":"9364595b-9b85-47dd-adb3-357db3855ea6","Type":"ContainerDied","Data":"5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d"} Dec 03 14:20:46 crc kubenswrapper[4690]: I1203 14:20:46.638352 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x849l" event={"ID":"795c8cf8-c96f-4bab-86d1-5881b53d40f9","Type":"ContainerStarted","Data":"b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3"} Dec 03 14:20:46 crc kubenswrapper[4690]: I1203 14:20:46.663903 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x849l" podStartSLOduration=2.632354521 podStartE2EDuration="8.663858293s" podCreationTimestamp="2025-12-03 14:20:38 +0000 UTC" firstStartedPulling="2025-12-03 14:20:39.528805642 +0000 UTC m=+4285.509726075" lastFinishedPulling="2025-12-03 14:20:45.560309414 +0000 UTC m=+4291.541229847" observedRunningTime="2025-12-03 14:20:46.653740377 +0000 UTC m=+4292.634660810" watchObservedRunningTime="2025-12-03 14:20:46.663858293 +0000 UTC m=+4292.644778716" Dec 03 14:20:46 crc kubenswrapper[4690]: I1203 14:20:46.825218 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:20:46 crc kubenswrapper[4690]: I1203 14:20:46.825271 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.295311 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-57qfs/must-gather-lzcsm"] Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.295748 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-57qfs/must-gather-lzcsm" podUID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerName="copy" containerID="cri-o://2d2b632a39484f16c71618a830f8238bbc2002337fe94ba6fef8f64b89a7c929" gracePeriod=2 Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.317770 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-57qfs/must-gather-lzcsm"] Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.651530 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-57qfs_must-gather-lzcsm_cf253854-f14d-4ea6-9f4b-6e988d5fb432/copy/0.log" Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.652626 4690 generic.go:334] "Generic (PLEG): container finished" podID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerID="2d2b632a39484f16c71618a830f8238bbc2002337fe94ba6fef8f64b89a7c929" exitCode=143 Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.656947 4690 generic.go:334] "Generic (PLEG): container finished" podID="9364595b-9b85-47dd-adb3-357db3855ea6" containerID="c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0" exitCode=0 Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.656989 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hcl6" event={"ID":"9364595b-9b85-47dd-adb3-357db3855ea6","Type":"ContainerDied","Data":"c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0"} Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.990761 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-57qfs_must-gather-lzcsm_cf253854-f14d-4ea6-9f4b-6e988d5fb432/copy/0.log" Dec 03 14:20:47 crc kubenswrapper[4690]: I1203 14:20:47.991677 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.030089 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf253854-f14d-4ea6-9f4b-6e988d5fb432-must-gather-output\") pod \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\" (UID: \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\") " Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.131394 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp5vh\" (UniqueName: \"kubernetes.io/projected/cf253854-f14d-4ea6-9f4b-6e988d5fb432-kube-api-access-lp5vh\") pod \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\" (UID: \"cf253854-f14d-4ea6-9f4b-6e988d5fb432\") " Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.139995 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf253854-f14d-4ea6-9f4b-6e988d5fb432-kube-api-access-lp5vh" (OuterVolumeSpecName: "kube-api-access-lp5vh") pod "cf253854-f14d-4ea6-9f4b-6e988d5fb432" (UID: "cf253854-f14d-4ea6-9f4b-6e988d5fb432"). InnerVolumeSpecName "kube-api-access-lp5vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.171991 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf253854-f14d-4ea6-9f4b-6e988d5fb432-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cf253854-f14d-4ea6-9f4b-6e988d5fb432" (UID: "cf253854-f14d-4ea6-9f4b-6e988d5fb432"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.233791 4690 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf253854-f14d-4ea6-9f4b-6e988d5fb432-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.233841 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp5vh\" (UniqueName: \"kubernetes.io/projected/cf253854-f14d-4ea6-9f4b-6e988d5fb432-kube-api-access-lp5vh\") on node \"crc\" DevicePath \"\"" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.327927 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" path="/var/lib/kubelet/pods/cf253854-f14d-4ea6-9f4b-6e988d5fb432/volumes" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.453089 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.453158 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.668475 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-57qfs_must-gather-lzcsm_cf253854-f14d-4ea6-9f4b-6e988d5fb432/copy/0.log" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.669555 4690 scope.go:117] "RemoveContainer" containerID="2d2b632a39484f16c71618a830f8238bbc2002337fe94ba6fef8f64b89a7c929" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.669719 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57qfs/must-gather-lzcsm" Dec 03 14:20:48 crc kubenswrapper[4690]: I1203 14:20:48.698144 4690 scope.go:117] "RemoveContainer" containerID="e343e8a0a0ab1358feb642c0c16e1dcc26de2fbbd7c0c36b563162cefec04d7b" Dec 03 14:20:49 crc kubenswrapper[4690]: I1203 14:20:49.516777 4690 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x849l" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="registry-server" probeResult="failure" output=< Dec 03 14:20:49 crc kubenswrapper[4690]: timeout: failed to connect service ":50051" within 1s Dec 03 14:20:49 crc kubenswrapper[4690]: > Dec 03 14:20:49 crc kubenswrapper[4690]: I1203 14:20:49.686076 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hcl6" event={"ID":"9364595b-9b85-47dd-adb3-357db3855ea6","Type":"ContainerStarted","Data":"5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110"} Dec 03 14:20:49 crc kubenswrapper[4690]: I1203 14:20:49.709984 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8hcl6" podStartSLOduration=5.811891966 podStartE2EDuration="8.709953571s" podCreationTimestamp="2025-12-03 14:20:41 +0000 UTC" firstStartedPulling="2025-12-03 14:20:45.628997781 +0000 UTC m=+4291.609918214" lastFinishedPulling="2025-12-03 14:20:48.527059386 +0000 UTC m=+4294.507979819" observedRunningTime="2025-12-03 14:20:49.707648143 +0000 UTC m=+4295.688568586" watchObservedRunningTime="2025-12-03 14:20:49.709953571 +0000 UTC m=+4295.690874004" Dec 03 14:20:51 crc kubenswrapper[4690]: I1203 14:20:51.645860 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:51 crc kubenswrapper[4690]: I1203 14:20:51.646279 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:51 crc kubenswrapper[4690]: I1203 14:20:51.694695 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:20:58 crc kubenswrapper[4690]: I1203 14:20:58.504502 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:58 crc kubenswrapper[4690]: I1203 14:20:58.558043 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:20:58 crc kubenswrapper[4690]: I1203 14:20:58.746639 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x849l"] Dec 03 14:20:59 crc kubenswrapper[4690]: I1203 14:20:59.880062 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x849l" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="registry-server" containerID="cri-o://b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3" gracePeriod=2 Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.390236 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.513034 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-utilities\") pod \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.513153 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7fb9\" (UniqueName: \"kubernetes.io/projected/795c8cf8-c96f-4bab-86d1-5881b53d40f9-kube-api-access-n7fb9\") pod \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.513413 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-catalog-content\") pod \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\" (UID: \"795c8cf8-c96f-4bab-86d1-5881b53d40f9\") " Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.516163 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-utilities" (OuterVolumeSpecName: "utilities") pod "795c8cf8-c96f-4bab-86d1-5881b53d40f9" (UID: "795c8cf8-c96f-4bab-86d1-5881b53d40f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.528162 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/795c8cf8-c96f-4bab-86d1-5881b53d40f9-kube-api-access-n7fb9" (OuterVolumeSpecName: "kube-api-access-n7fb9") pod "795c8cf8-c96f-4bab-86d1-5881b53d40f9" (UID: "795c8cf8-c96f-4bab-86d1-5881b53d40f9"). InnerVolumeSpecName "kube-api-access-n7fb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.616179 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.616231 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7fb9\" (UniqueName: \"kubernetes.io/projected/795c8cf8-c96f-4bab-86d1-5881b53d40f9-kube-api-access-n7fb9\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.653635 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "795c8cf8-c96f-4bab-86d1-5881b53d40f9" (UID: "795c8cf8-c96f-4bab-86d1-5881b53d40f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.718538 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/795c8cf8-c96f-4bab-86d1-5881b53d40f9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.894003 4690 generic.go:334] "Generic (PLEG): container finished" podID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerID="b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3" exitCode=0 Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.894181 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x849l" event={"ID":"795c8cf8-c96f-4bab-86d1-5881b53d40f9","Type":"ContainerDied","Data":"b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3"} Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.894419 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x849l" event={"ID":"795c8cf8-c96f-4bab-86d1-5881b53d40f9","Type":"ContainerDied","Data":"3efe696d091068fab500c3b1fdc687f0355fc8c91f2628126bf861cabd747b02"} Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.894447 4690 scope.go:117] "RemoveContainer" containerID="b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.894210 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x849l" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.923654 4690 scope.go:117] "RemoveContainer" containerID="4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6" Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.938602 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x849l"] Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.948666 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x849l"] Dec 03 14:21:00 crc kubenswrapper[4690]: I1203 14:21:00.964191 4690 scope.go:117] "RemoveContainer" containerID="928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340" Dec 03 14:21:01 crc kubenswrapper[4690]: I1203 14:21:01.024463 4690 scope.go:117] "RemoveContainer" containerID="b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3" Dec 03 14:21:01 crc kubenswrapper[4690]: E1203 14:21:01.024942 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3\": container with ID starting with b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3 not found: ID does not exist" containerID="b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3" Dec 03 14:21:01 crc kubenswrapper[4690]: I1203 14:21:01.024990 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3"} err="failed to get container status \"b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3\": rpc error: code = NotFound desc = could not find container \"b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3\": container with ID starting with b152e2f1ca9701f8fdec9e1f5ec7e8721383813ee4b30426d9e6d1b29de0a3f3 not found: ID does not exist" Dec 03 14:21:01 crc kubenswrapper[4690]: I1203 14:21:01.025018 4690 scope.go:117] "RemoveContainer" containerID="4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6" Dec 03 14:21:01 crc kubenswrapper[4690]: E1203 14:21:01.025320 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6\": container with ID starting with 4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6 not found: ID does not exist" containerID="4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6" Dec 03 14:21:01 crc kubenswrapper[4690]: I1203 14:21:01.025381 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6"} err="failed to get container status \"4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6\": rpc error: code = NotFound desc = could not find container \"4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6\": container with ID starting with 4a319e392090f2fc83486659eb7853c4a2f56733051ddbb47778bc20bca2afe6 not found: ID does not exist" Dec 03 14:21:01 crc kubenswrapper[4690]: I1203 14:21:01.025425 4690 scope.go:117] "RemoveContainer" containerID="928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340" Dec 03 14:21:01 crc kubenswrapper[4690]: E1203 14:21:01.025725 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340\": container with ID starting with 928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340 not found: ID does not exist" containerID="928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340" Dec 03 14:21:01 crc kubenswrapper[4690]: I1203 14:21:01.025754 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340"} err="failed to get container status \"928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340\": rpc error: code = NotFound desc = could not find container \"928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340\": container with ID starting with 928e6ce12c72a5366ce4cf7565627c5f3419c143957ebb9a5b4d77c99e705340 not found: ID does not exist" Dec 03 14:21:01 crc kubenswrapper[4690]: I1203 14:21:01.699589 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:21:02 crc kubenswrapper[4690]: I1203 14:21:02.327921 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" path="/var/lib/kubelet/pods/795c8cf8-c96f-4bab-86d1-5881b53d40f9/volumes" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.141216 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hcl6"] Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.141791 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8hcl6" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" containerName="registry-server" containerID="cri-o://5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110" gracePeriod=2 Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.622598 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.787385 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg6md\" (UniqueName: \"kubernetes.io/projected/9364595b-9b85-47dd-adb3-357db3855ea6-kube-api-access-cg6md\") pod \"9364595b-9b85-47dd-adb3-357db3855ea6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.787735 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-utilities\") pod \"9364595b-9b85-47dd-adb3-357db3855ea6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.787984 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-catalog-content\") pod \"9364595b-9b85-47dd-adb3-357db3855ea6\" (UID: \"9364595b-9b85-47dd-adb3-357db3855ea6\") " Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.789576 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-utilities" (OuterVolumeSpecName: "utilities") pod "9364595b-9b85-47dd-adb3-357db3855ea6" (UID: "9364595b-9b85-47dd-adb3-357db3855ea6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.800184 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9364595b-9b85-47dd-adb3-357db3855ea6-kube-api-access-cg6md" (OuterVolumeSpecName: "kube-api-access-cg6md") pod "9364595b-9b85-47dd-adb3-357db3855ea6" (UID: "9364595b-9b85-47dd-adb3-357db3855ea6"). InnerVolumeSpecName "kube-api-access-cg6md". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.808296 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9364595b-9b85-47dd-adb3-357db3855ea6" (UID: "9364595b-9b85-47dd-adb3-357db3855ea6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.891194 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.891238 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg6md\" (UniqueName: \"kubernetes.io/projected/9364595b-9b85-47dd-adb3-357db3855ea6-kube-api-access-cg6md\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.891256 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9364595b-9b85-47dd-adb3-357db3855ea6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.930134 4690 generic.go:334] "Generic (PLEG): container finished" podID="9364595b-9b85-47dd-adb3-357db3855ea6" containerID="5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110" exitCode=0 Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.930187 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hcl6" event={"ID":"9364595b-9b85-47dd-adb3-357db3855ea6","Type":"ContainerDied","Data":"5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110"} Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.930221 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hcl6" event={"ID":"9364595b-9b85-47dd-adb3-357db3855ea6","Type":"ContainerDied","Data":"ea7ed38b933522142abdac3a949798914983132ad6367c1d484e00c32a466aab"} Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.930245 4690 scope.go:117] "RemoveContainer" containerID="5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.930368 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hcl6" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.955541 4690 scope.go:117] "RemoveContainer" containerID="c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0" Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.970840 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hcl6"] Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.981527 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hcl6"] Dec 03 14:21:03 crc kubenswrapper[4690]: I1203 14:21:03.992024 4690 scope.go:117] "RemoveContainer" containerID="5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d" Dec 03 14:21:04 crc kubenswrapper[4690]: I1203 14:21:04.041901 4690 scope.go:117] "RemoveContainer" containerID="5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110" Dec 03 14:21:04 crc kubenswrapper[4690]: E1203 14:21:04.042469 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110\": container with ID starting with 5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110 not found: ID does not exist" containerID="5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110" Dec 03 14:21:04 crc kubenswrapper[4690]: I1203 14:21:04.042517 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110"} err="failed to get container status \"5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110\": rpc error: code = NotFound desc = could not find container \"5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110\": container with ID starting with 5dfae1dcd68d5f8912a59d4135fb21dd3c846da9293184e1cda063b73c768110 not found: ID does not exist" Dec 03 14:21:04 crc kubenswrapper[4690]: I1203 14:21:04.042548 4690 scope.go:117] "RemoveContainer" containerID="c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0" Dec 03 14:21:04 crc kubenswrapper[4690]: E1203 14:21:04.043256 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0\": container with ID starting with c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0 not found: ID does not exist" containerID="c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0" Dec 03 14:21:04 crc kubenswrapper[4690]: I1203 14:21:04.043294 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0"} err="failed to get container status \"c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0\": rpc error: code = NotFound desc = could not find container \"c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0\": container with ID starting with c738dee4fa6b0cf8c4305f3ade11464ccf31e81fcbf5826832f0607adc9d58e0 not found: ID does not exist" Dec 03 14:21:04 crc kubenswrapper[4690]: I1203 14:21:04.043317 4690 scope.go:117] "RemoveContainer" containerID="5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d" Dec 03 14:21:04 crc kubenswrapper[4690]: E1203 14:21:04.043604 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d\": container with ID starting with 5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d not found: ID does not exist" containerID="5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d" Dec 03 14:21:04 crc kubenswrapper[4690]: I1203 14:21:04.043638 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d"} err="failed to get container status \"5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d\": rpc error: code = NotFound desc = could not find container \"5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d\": container with ID starting with 5f6f67445dd65f87e60a1472c7c68a1989f7c543ac414a55618088b1bce56f5d not found: ID does not exist" Dec 03 14:21:04 crc kubenswrapper[4690]: I1203 14:21:04.336929 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" path="/var/lib/kubelet/pods/9364595b-9b85-47dd-adb3-357db3855ea6/volumes" Dec 03 14:21:16 crc kubenswrapper[4690]: I1203 14:21:16.823956 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:21:16 crc kubenswrapper[4690]: I1203 14:21:16.824660 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:21:46 crc kubenswrapper[4690]: I1203 14:21:46.823498 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:21:46 crc kubenswrapper[4690]: I1203 14:21:46.825055 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:21:46 crc kubenswrapper[4690]: I1203 14:21:46.825125 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 14:21:46 crc kubenswrapper[4690]: I1203 14:21:46.826037 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2c14307ea04a24716f0f1945ac9b63fea01b9e0b074627d7758c3e1e8b30cd4"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:21:46 crc kubenswrapper[4690]: I1203 14:21:46.826104 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://c2c14307ea04a24716f0f1945ac9b63fea01b9e0b074627d7758c3e1e8b30cd4" gracePeriod=600 Dec 03 14:21:47 crc kubenswrapper[4690]: I1203 14:21:47.362168 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="c2c14307ea04a24716f0f1945ac9b63fea01b9e0b074627d7758c3e1e8b30cd4" exitCode=0 Dec 03 14:21:47 crc kubenswrapper[4690]: I1203 14:21:47.362230 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"c2c14307ea04a24716f0f1945ac9b63fea01b9e0b074627d7758c3e1e8b30cd4"} Dec 03 14:21:47 crc kubenswrapper[4690]: I1203 14:21:47.362661 4690 scope.go:117] "RemoveContainer" containerID="d848116b4b1e834bcf94069d6b71fad903b9d04aa105149ca9368084993817af" Dec 03 14:21:48 crc kubenswrapper[4690]: I1203 14:21:48.374748 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb"} Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.307052 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksl6l/must-gather-vwxpt"] Dec 03 14:24:01 crc kubenswrapper[4690]: E1203 14:24:01.308096 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="registry-server" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308112 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="registry-server" Dec 03 14:24:01 crc kubenswrapper[4690]: E1203 14:24:01.308132 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerName="copy" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308138 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerName="copy" Dec 03 14:24:01 crc kubenswrapper[4690]: E1203 14:24:01.308151 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="extract-utilities" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308158 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="extract-utilities" Dec 03 14:24:01 crc kubenswrapper[4690]: E1203 14:24:01.308175 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" containerName="extract-content" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308182 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" containerName="extract-content" Dec 03 14:24:01 crc kubenswrapper[4690]: E1203 14:24:01.308202 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerName="gather" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308213 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerName="gather" Dec 03 14:24:01 crc kubenswrapper[4690]: E1203 14:24:01.308227 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="extract-content" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308232 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="extract-content" Dec 03 14:24:01 crc kubenswrapper[4690]: E1203 14:24:01.308242 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" containerName="registry-server" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308247 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" containerName="registry-server" Dec 03 14:24:01 crc kubenswrapper[4690]: E1203 14:24:01.308257 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" containerName="extract-utilities" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308263 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" containerName="extract-utilities" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308478 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="9364595b-9b85-47dd-adb3-357db3855ea6" containerName="registry-server" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308492 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerName="copy" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308501 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf253854-f14d-4ea6-9f4b-6e988d5fb432" containerName="gather" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.308509 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="795c8cf8-c96f-4bab-86d1-5881b53d40f9" containerName="registry-server" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.309678 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.317706 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ksl6l"/"openshift-service-ca.crt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.319351 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ksl6l"/"kube-root-ca.crt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.325139 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ksl6l"/"default-dockercfg-hp5ns" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.342479 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ksl6l/must-gather-vwxpt"] Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.354022 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b36061d5-2fe4-4c05-8c07-02674c7337ec-must-gather-output\") pod \"must-gather-vwxpt\" (UID: \"b36061d5-2fe4-4c05-8c07-02674c7337ec\") " pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.354118 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlqfm\" (UniqueName: \"kubernetes.io/projected/b36061d5-2fe4-4c05-8c07-02674c7337ec-kube-api-access-hlqfm\") pod \"must-gather-vwxpt\" (UID: \"b36061d5-2fe4-4c05-8c07-02674c7337ec\") " pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.456252 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b36061d5-2fe4-4c05-8c07-02674c7337ec-must-gather-output\") pod \"must-gather-vwxpt\" (UID: \"b36061d5-2fe4-4c05-8c07-02674c7337ec\") " pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.456345 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlqfm\" (UniqueName: \"kubernetes.io/projected/b36061d5-2fe4-4c05-8c07-02674c7337ec-kube-api-access-hlqfm\") pod \"must-gather-vwxpt\" (UID: \"b36061d5-2fe4-4c05-8c07-02674c7337ec\") " pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.456891 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b36061d5-2fe4-4c05-8c07-02674c7337ec-must-gather-output\") pod \"must-gather-vwxpt\" (UID: \"b36061d5-2fe4-4c05-8c07-02674c7337ec\") " pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.490502 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlqfm\" (UniqueName: \"kubernetes.io/projected/b36061d5-2fe4-4c05-8c07-02674c7337ec-kube-api-access-hlqfm\") pod \"must-gather-vwxpt\" (UID: \"b36061d5-2fe4-4c05-8c07-02674c7337ec\") " pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:24:01 crc kubenswrapper[4690]: I1203 14:24:01.636457 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:24:02 crc kubenswrapper[4690]: I1203 14:24:02.165202 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ksl6l/must-gather-vwxpt"] Dec 03 14:24:02 crc kubenswrapper[4690]: I1203 14:24:02.939931 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" event={"ID":"b36061d5-2fe4-4c05-8c07-02674c7337ec","Type":"ContainerStarted","Data":"bad2814e8c4e4ee3fd7672d739a950136cdbd5c231e895d35c1819539b142cf0"} Dec 03 14:24:03 crc kubenswrapper[4690]: I1203 14:24:03.952333 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" event={"ID":"b36061d5-2fe4-4c05-8c07-02674c7337ec","Type":"ContainerStarted","Data":"90e73552df83a70985013aef2b413554dec9f4d024b05b689b42fcf4a4c5af54"} Dec 03 14:24:03 crc kubenswrapper[4690]: I1203 14:24:03.952722 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" event={"ID":"b36061d5-2fe4-4c05-8c07-02674c7337ec","Type":"ContainerStarted","Data":"887da7cca4071f6ea28acd15a8733ad0bbd8c38940c481168c8f804c5209a62b"} Dec 03 14:24:03 crc kubenswrapper[4690]: I1203 14:24:03.977849 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" podStartSLOduration=2.977802534 podStartE2EDuration="2.977802534s" podCreationTimestamp="2025-12-03 14:24:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:24:03.965478291 +0000 UTC m=+4489.946398724" watchObservedRunningTime="2025-12-03 14:24:03.977802534 +0000 UTC m=+4489.958722967" Dec 03 14:24:06 crc kubenswrapper[4690]: I1203 14:24:06.929662 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-schpc"] Dec 03 14:24:06 crc kubenswrapper[4690]: I1203 14:24:06.931659 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.084727 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rnqg\" (UniqueName: \"kubernetes.io/projected/b918b51a-c60d-4ada-a971-d132f7a18c77-kube-api-access-6rnqg\") pod \"crc-debug-schpc\" (UID: \"b918b51a-c60d-4ada-a971-d132f7a18c77\") " pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.084953 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b918b51a-c60d-4ada-a971-d132f7a18c77-host\") pod \"crc-debug-schpc\" (UID: \"b918b51a-c60d-4ada-a971-d132f7a18c77\") " pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.186375 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b918b51a-c60d-4ada-a971-d132f7a18c77-host\") pod \"crc-debug-schpc\" (UID: \"b918b51a-c60d-4ada-a971-d132f7a18c77\") " pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.186798 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rnqg\" (UniqueName: \"kubernetes.io/projected/b918b51a-c60d-4ada-a971-d132f7a18c77-kube-api-access-6rnqg\") pod \"crc-debug-schpc\" (UID: \"b918b51a-c60d-4ada-a971-d132f7a18c77\") " pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.186482 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b918b51a-c60d-4ada-a971-d132f7a18c77-host\") pod \"crc-debug-schpc\" (UID: \"b918b51a-c60d-4ada-a971-d132f7a18c77\") " pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.209378 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rnqg\" (UniqueName: \"kubernetes.io/projected/b918b51a-c60d-4ada-a971-d132f7a18c77-kube-api-access-6rnqg\") pod \"crc-debug-schpc\" (UID: \"b918b51a-c60d-4ada-a971-d132f7a18c77\") " pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.252879 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:07 crc kubenswrapper[4690]: W1203 14:24:07.289629 4690 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb918b51a_c60d_4ada_a971_d132f7a18c77.slice/crio-cc516a045de290dabdd1bf194687b12e9845a2f706888f0ee905107b24a67d5f WatchSource:0}: Error finding container cc516a045de290dabdd1bf194687b12e9845a2f706888f0ee905107b24a67d5f: Status 404 returned error can't find the container with id cc516a045de290dabdd1bf194687b12e9845a2f706888f0ee905107b24a67d5f Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.992642 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/crc-debug-schpc" event={"ID":"b918b51a-c60d-4ada-a971-d132f7a18c77","Type":"ContainerStarted","Data":"169c8b15637f1ff2e65a6678b6f409d618ad651e29c1789f007e16c50208f91e"} Dec 03 14:24:07 crc kubenswrapper[4690]: I1203 14:24:07.993044 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/crc-debug-schpc" event={"ID":"b918b51a-c60d-4ada-a971-d132f7a18c77","Type":"ContainerStarted","Data":"cc516a045de290dabdd1bf194687b12e9845a2f706888f0ee905107b24a67d5f"} Dec 03 14:24:08 crc kubenswrapper[4690]: I1203 14:24:08.019751 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ksl6l/crc-debug-schpc" podStartSLOduration=2.019727356 podStartE2EDuration="2.019727356s" podCreationTimestamp="2025-12-03 14:24:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:24:08.009293411 +0000 UTC m=+4493.990213864" watchObservedRunningTime="2025-12-03 14:24:08.019727356 +0000 UTC m=+4494.000647789" Dec 03 14:24:16 crc kubenswrapper[4690]: I1203 14:24:16.824083 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:24:16 crc kubenswrapper[4690]: I1203 14:24:16.824734 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:24:36 crc kubenswrapper[4690]: I1203 14:24:36.899878 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-696775c5d7-q6gwf" podUID="9975bb19-d67e-4a25-a69f-8c1e37653961" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 03 14:24:46 crc kubenswrapper[4690]: I1203 14:24:46.823451 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:24:46 crc kubenswrapper[4690]: I1203 14:24:46.824089 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:24:49 crc kubenswrapper[4690]: I1203 14:24:49.434549 4690 generic.go:334] "Generic (PLEG): container finished" podID="b918b51a-c60d-4ada-a971-d132f7a18c77" containerID="169c8b15637f1ff2e65a6678b6f409d618ad651e29c1789f007e16c50208f91e" exitCode=0 Dec 03 14:24:49 crc kubenswrapper[4690]: I1203 14:24:49.434619 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/crc-debug-schpc" event={"ID":"b918b51a-c60d-4ada-a971-d132f7a18c77","Type":"ContainerDied","Data":"169c8b15637f1ff2e65a6678b6f409d618ad651e29c1789f007e16c50208f91e"} Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.679650 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.716400 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-schpc"] Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.725319 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-schpc"] Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.778540 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b918b51a-c60d-4ada-a971-d132f7a18c77-host\") pod \"b918b51a-c60d-4ada-a971-d132f7a18c77\" (UID: \"b918b51a-c60d-4ada-a971-d132f7a18c77\") " Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.778646 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rnqg\" (UniqueName: \"kubernetes.io/projected/b918b51a-c60d-4ada-a971-d132f7a18c77-kube-api-access-6rnqg\") pod \"b918b51a-c60d-4ada-a971-d132f7a18c77\" (UID: \"b918b51a-c60d-4ada-a971-d132f7a18c77\") " Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.778764 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b918b51a-c60d-4ada-a971-d132f7a18c77-host" (OuterVolumeSpecName: "host") pod "b918b51a-c60d-4ada-a971-d132f7a18c77" (UID: "b918b51a-c60d-4ada-a971-d132f7a18c77"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.779193 4690 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b918b51a-c60d-4ada-a971-d132f7a18c77-host\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.786139 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b918b51a-c60d-4ada-a971-d132f7a18c77-kube-api-access-6rnqg" (OuterVolumeSpecName: "kube-api-access-6rnqg") pod "b918b51a-c60d-4ada-a971-d132f7a18c77" (UID: "b918b51a-c60d-4ada-a971-d132f7a18c77"). InnerVolumeSpecName "kube-api-access-6rnqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:50 crc kubenswrapper[4690]: I1203 14:24:50.881544 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rnqg\" (UniqueName: \"kubernetes.io/projected/b918b51a-c60d-4ada-a971-d132f7a18c77-kube-api-access-6rnqg\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:51 crc kubenswrapper[4690]: I1203 14:24:51.460375 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc516a045de290dabdd1bf194687b12e9845a2f706888f0ee905107b24a67d5f" Dec 03 14:24:51 crc kubenswrapper[4690]: I1203 14:24:51.460657 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-schpc" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.005017 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-xcd2j"] Dec 03 14:24:52 crc kubenswrapper[4690]: E1203 14:24:52.005618 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b918b51a-c60d-4ada-a971-d132f7a18c77" containerName="container-00" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.005637 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b918b51a-c60d-4ada-a971-d132f7a18c77" containerName="container-00" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.005915 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b918b51a-c60d-4ada-a971-d132f7a18c77" containerName="container-00" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.006925 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.107651 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng7qg\" (UniqueName: \"kubernetes.io/projected/940816e8-d68c-481b-abe0-feeb6ca55255-kube-api-access-ng7qg\") pod \"crc-debug-xcd2j\" (UID: \"940816e8-d68c-481b-abe0-feeb6ca55255\") " pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.107759 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940816e8-d68c-481b-abe0-feeb6ca55255-host\") pod \"crc-debug-xcd2j\" (UID: \"940816e8-d68c-481b-abe0-feeb6ca55255\") " pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.210029 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng7qg\" (UniqueName: \"kubernetes.io/projected/940816e8-d68c-481b-abe0-feeb6ca55255-kube-api-access-ng7qg\") pod \"crc-debug-xcd2j\" (UID: \"940816e8-d68c-481b-abe0-feeb6ca55255\") " pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.210124 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940816e8-d68c-481b-abe0-feeb6ca55255-host\") pod \"crc-debug-xcd2j\" (UID: \"940816e8-d68c-481b-abe0-feeb6ca55255\") " pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.210253 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940816e8-d68c-481b-abe0-feeb6ca55255-host\") pod \"crc-debug-xcd2j\" (UID: \"940816e8-d68c-481b-abe0-feeb6ca55255\") " pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.231120 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng7qg\" (UniqueName: \"kubernetes.io/projected/940816e8-d68c-481b-abe0-feeb6ca55255-kube-api-access-ng7qg\") pod \"crc-debug-xcd2j\" (UID: \"940816e8-d68c-481b-abe0-feeb6ca55255\") " pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.330677 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b918b51a-c60d-4ada-a971-d132f7a18c77" path="/var/lib/kubelet/pods/b918b51a-c60d-4ada-a971-d132f7a18c77/volumes" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.337954 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:52 crc kubenswrapper[4690]: I1203 14:24:52.501814 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" event={"ID":"940816e8-d68c-481b-abe0-feeb6ca55255","Type":"ContainerStarted","Data":"9e9732dbb3f0b871c63455a5e970c78dfaf1ea09b6c138ff32c75f227408d78d"} Dec 03 14:24:53 crc kubenswrapper[4690]: I1203 14:24:53.517094 4690 generic.go:334] "Generic (PLEG): container finished" podID="940816e8-d68c-481b-abe0-feeb6ca55255" containerID="1b926664028e64d3be77ae616c1974ca9342c7752c896e3014b49134ec814ffd" exitCode=0 Dec 03 14:24:53 crc kubenswrapper[4690]: I1203 14:24:53.517189 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" event={"ID":"940816e8-d68c-481b-abe0-feeb6ca55255","Type":"ContainerDied","Data":"1b926664028e64d3be77ae616c1974ca9342c7752c896e3014b49134ec814ffd"} Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.642813 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.710449 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-xcd2j"] Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.719776 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-xcd2j"] Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.765653 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940816e8-d68c-481b-abe0-feeb6ca55255-host\") pod \"940816e8-d68c-481b-abe0-feeb6ca55255\" (UID: \"940816e8-d68c-481b-abe0-feeb6ca55255\") " Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.765798 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/940816e8-d68c-481b-abe0-feeb6ca55255-host" (OuterVolumeSpecName: "host") pod "940816e8-d68c-481b-abe0-feeb6ca55255" (UID: "940816e8-d68c-481b-abe0-feeb6ca55255"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.765975 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng7qg\" (UniqueName: \"kubernetes.io/projected/940816e8-d68c-481b-abe0-feeb6ca55255-kube-api-access-ng7qg\") pod \"940816e8-d68c-481b-abe0-feeb6ca55255\" (UID: \"940816e8-d68c-481b-abe0-feeb6ca55255\") " Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.766707 4690 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940816e8-d68c-481b-abe0-feeb6ca55255-host\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.777269 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/940816e8-d68c-481b-abe0-feeb6ca55255-kube-api-access-ng7qg" (OuterVolumeSpecName: "kube-api-access-ng7qg") pod "940816e8-d68c-481b-abe0-feeb6ca55255" (UID: "940816e8-d68c-481b-abe0-feeb6ca55255"). InnerVolumeSpecName "kube-api-access-ng7qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:54 crc kubenswrapper[4690]: I1203 14:24:54.868550 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng7qg\" (UniqueName: \"kubernetes.io/projected/940816e8-d68c-481b-abe0-feeb6ca55255-kube-api-access-ng7qg\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:55 crc kubenswrapper[4690]: I1203 14:24:55.555767 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e9732dbb3f0b871c63455a5e970c78dfaf1ea09b6c138ff32c75f227408d78d" Dec 03 14:24:55 crc kubenswrapper[4690]: I1203 14:24:55.555902 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-xcd2j" Dec 03 14:24:55 crc kubenswrapper[4690]: I1203 14:24:55.881971 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-4wx5x"] Dec 03 14:24:55 crc kubenswrapper[4690]: E1203 14:24:55.882980 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940816e8-d68c-481b-abe0-feeb6ca55255" containerName="container-00" Dec 03 14:24:55 crc kubenswrapper[4690]: I1203 14:24:55.882998 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="940816e8-d68c-481b-abe0-feeb6ca55255" containerName="container-00" Dec 03 14:24:55 crc kubenswrapper[4690]: I1203 14:24:55.883243 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="940816e8-d68c-481b-abe0-feeb6ca55255" containerName="container-00" Dec 03 14:24:55 crc kubenswrapper[4690]: I1203 14:24:55.884252 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.005051 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pkmj\" (UniqueName: \"kubernetes.io/projected/446b00f8-b816-4daa-903e-7f370be05883-kube-api-access-9pkmj\") pod \"crc-debug-4wx5x\" (UID: \"446b00f8-b816-4daa-903e-7f370be05883\") " pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.005103 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/446b00f8-b816-4daa-903e-7f370be05883-host\") pod \"crc-debug-4wx5x\" (UID: \"446b00f8-b816-4daa-903e-7f370be05883\") " pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.107509 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pkmj\" (UniqueName: \"kubernetes.io/projected/446b00f8-b816-4daa-903e-7f370be05883-kube-api-access-9pkmj\") pod \"crc-debug-4wx5x\" (UID: \"446b00f8-b816-4daa-903e-7f370be05883\") " pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.107600 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/446b00f8-b816-4daa-903e-7f370be05883-host\") pod \"crc-debug-4wx5x\" (UID: \"446b00f8-b816-4daa-903e-7f370be05883\") " pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.107896 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/446b00f8-b816-4daa-903e-7f370be05883-host\") pod \"crc-debug-4wx5x\" (UID: \"446b00f8-b816-4daa-903e-7f370be05883\") " pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.129012 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pkmj\" (UniqueName: \"kubernetes.io/projected/446b00f8-b816-4daa-903e-7f370be05883-kube-api-access-9pkmj\") pod \"crc-debug-4wx5x\" (UID: \"446b00f8-b816-4daa-903e-7f370be05883\") " pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.203002 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.329401 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="940816e8-d68c-481b-abe0-feeb6ca55255" path="/var/lib/kubelet/pods/940816e8-d68c-481b-abe0-feeb6ca55255/volumes" Dec 03 14:24:56 crc kubenswrapper[4690]: I1203 14:24:56.568669 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" event={"ID":"446b00f8-b816-4daa-903e-7f370be05883","Type":"ContainerStarted","Data":"20c1a9fbd012fa6d5d21697d692f2104e65b7911d17eaf30326a612e1c5a28e1"} Dec 03 14:24:57 crc kubenswrapper[4690]: I1203 14:24:57.582170 4690 generic.go:334] "Generic (PLEG): container finished" podID="446b00f8-b816-4daa-903e-7f370be05883" containerID="d1385b7f08543090040619e8890b339781ce50a55c2d29aa22f8b6e9236db843" exitCode=0 Dec 03 14:24:57 crc kubenswrapper[4690]: I1203 14:24:57.582543 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" event={"ID":"446b00f8-b816-4daa-903e-7f370be05883","Type":"ContainerDied","Data":"d1385b7f08543090040619e8890b339781ce50a55c2d29aa22f8b6e9236db843"} Dec 03 14:24:57 crc kubenswrapper[4690]: I1203 14:24:57.621688 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-4wx5x"] Dec 03 14:24:57 crc kubenswrapper[4690]: I1203 14:24:57.630705 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksl6l/crc-debug-4wx5x"] Dec 03 14:24:58 crc kubenswrapper[4690]: I1203 14:24:58.706420 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:24:58 crc kubenswrapper[4690]: I1203 14:24:58.787360 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/446b00f8-b816-4daa-903e-7f370be05883-host\") pod \"446b00f8-b816-4daa-903e-7f370be05883\" (UID: \"446b00f8-b816-4daa-903e-7f370be05883\") " Dec 03 14:24:58 crc kubenswrapper[4690]: I1203 14:24:58.787492 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446b00f8-b816-4daa-903e-7f370be05883-host" (OuterVolumeSpecName: "host") pod "446b00f8-b816-4daa-903e-7f370be05883" (UID: "446b00f8-b816-4daa-903e-7f370be05883"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:24:58 crc kubenswrapper[4690]: I1203 14:24:58.787538 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pkmj\" (UniqueName: \"kubernetes.io/projected/446b00f8-b816-4daa-903e-7f370be05883-kube-api-access-9pkmj\") pod \"446b00f8-b816-4daa-903e-7f370be05883\" (UID: \"446b00f8-b816-4daa-903e-7f370be05883\") " Dec 03 14:24:58 crc kubenswrapper[4690]: I1203 14:24:58.788311 4690 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/446b00f8-b816-4daa-903e-7f370be05883-host\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:58 crc kubenswrapper[4690]: I1203 14:24:58.800694 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/446b00f8-b816-4daa-903e-7f370be05883-kube-api-access-9pkmj" (OuterVolumeSpecName: "kube-api-access-9pkmj") pod "446b00f8-b816-4daa-903e-7f370be05883" (UID: "446b00f8-b816-4daa-903e-7f370be05883"). InnerVolumeSpecName "kube-api-access-9pkmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:58 crc kubenswrapper[4690]: I1203 14:24:58.890363 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pkmj\" (UniqueName: \"kubernetes.io/projected/446b00f8-b816-4daa-903e-7f370be05883-kube-api-access-9pkmj\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:59 crc kubenswrapper[4690]: I1203 14:24:59.604770 4690 scope.go:117] "RemoveContainer" containerID="d1385b7f08543090040619e8890b339781ce50a55c2d29aa22f8b6e9236db843" Dec 03 14:24:59 crc kubenswrapper[4690]: I1203 14:24:59.604831 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/crc-debug-4wx5x" Dec 03 14:25:00 crc kubenswrapper[4690]: I1203 14:25:00.328725 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="446b00f8-b816-4daa-903e-7f370be05883" path="/var/lib/kubelet/pods/446b00f8-b816-4daa-903e-7f370be05883/volumes" Dec 03 14:25:16 crc kubenswrapper[4690]: I1203 14:25:16.824920 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:25:16 crc kubenswrapper[4690]: I1203 14:25:16.825568 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:25:16 crc kubenswrapper[4690]: I1203 14:25:16.825625 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 14:25:16 crc kubenswrapper[4690]: I1203 14:25:16.826522 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:25:16 crc kubenswrapper[4690]: I1203 14:25:16.826584 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" gracePeriod=600 Dec 03 14:25:16 crc kubenswrapper[4690]: E1203 14:25:16.958474 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:25:17 crc kubenswrapper[4690]: I1203 14:25:17.835260 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" exitCode=0 Dec 03 14:25:17 crc kubenswrapper[4690]: I1203 14:25:17.835336 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb"} Dec 03 14:25:17 crc kubenswrapper[4690]: I1203 14:25:17.835699 4690 scope.go:117] "RemoveContainer" containerID="c2c14307ea04a24716f0f1945ac9b63fea01b9e0b074627d7758c3e1e8b30cd4" Dec 03 14:25:17 crc kubenswrapper[4690]: I1203 14:25:17.836852 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:25:17 crc kubenswrapper[4690]: E1203 14:25:17.837241 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:25:27 crc kubenswrapper[4690]: I1203 14:25:27.267252 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d8d5ff786-rbh82_1a74f0f7-bcc8-483f-ad5a-2df09e7f462d/barbican-api/0.log" Dec 03 14:25:27 crc kubenswrapper[4690]: I1203 14:25:27.510209 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d8d5ff786-rbh82_1a74f0f7-bcc8-483f-ad5a-2df09e7f462d/barbican-api-log/0.log" Dec 03 14:25:27 crc kubenswrapper[4690]: I1203 14:25:27.586589 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-98c9dfb78-qjztd_790eb625-6e7d-4467-afef-4073973889b2/barbican-keystone-listener/0.log" Dec 03 14:25:27 crc kubenswrapper[4690]: I1203 14:25:27.788461 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-98c9dfb78-qjztd_790eb625-6e7d-4467-afef-4073973889b2/barbican-keystone-listener-log/0.log" Dec 03 14:25:27 crc kubenswrapper[4690]: I1203 14:25:27.822639 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7c7c4b9c77-f54ln_f7a19182-b617-47e3-bc7f-db372257b984/barbican-worker/0.log" Dec 03 14:25:27 crc kubenswrapper[4690]: I1203 14:25:27.841830 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7c7c4b9c77-f54ln_f7a19182-b617-47e3-bc7f-db372257b984/barbican-worker-log/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.147553 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-nvbnl_ecded0ed-1510-4ac8-80d0-dd7b06c8446e/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.177359 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6330b5b-9599-49b6-89fa-b890374a7830/ceilometer-central-agent/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.400291 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6330b5b-9599-49b6-89fa-b890374a7830/sg-core/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.438152 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6330b5b-9599-49b6-89fa-b890374a7830/proxy-httpd/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.458948 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6330b5b-9599-49b6-89fa-b890374a7830/ceilometer-notification-agent/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.718345 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d33ca8c1-d71d-4624-a362-f845fda14976/cinder-api/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.733788 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d33ca8c1-d71d-4624-a362-f845fda14976/cinder-api-log/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.808678 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fa834771-9deb-40da-b5ae-4f8d60667d2b/cinder-scheduler/0.log" Dec 03 14:25:28 crc kubenswrapper[4690]: I1203 14:25:28.987480 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fa834771-9deb-40da-b5ae-4f8d60667d2b/probe/0.log" Dec 03 14:25:29 crc kubenswrapper[4690]: I1203 14:25:29.092187 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-p55xh_69f0b3dc-dc18-4a71-ad0c-a5023db9ca96/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:29 crc kubenswrapper[4690]: I1203 14:25:29.530569 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-77rvl_ede30151-25e0-47e4-98b4-bf7ca4d55255/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:29 crc kubenswrapper[4690]: I1203 14:25:29.622952 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-wn8bx_fa48f557-98dd-49ca-a4a6-89b196b77b15/init/0.log" Dec 03 14:25:29 crc kubenswrapper[4690]: I1203 14:25:29.909996 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-wn8bx_fa48f557-98dd-49ca-a4a6-89b196b77b15/init/0.log" Dec 03 14:25:29 crc kubenswrapper[4690]: I1203 14:25:29.925482 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-wn8bx_fa48f557-98dd-49ca-a4a6-89b196b77b15/dnsmasq-dns/0.log" Dec 03 14:25:29 crc kubenswrapper[4690]: I1203 14:25:29.929253 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pksbh_2f30c127-ad2e-4154-a344-61a21b57b3d0/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:30 crc kubenswrapper[4690]: I1203 14:25:30.141075 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ada5ac2f-e8b5-4437-b639-183fe2bf3c09/glance-log/0.log" Dec 03 14:25:30 crc kubenswrapper[4690]: I1203 14:25:30.156967 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ada5ac2f-e8b5-4437-b639-183fe2bf3c09/glance-httpd/0.log" Dec 03 14:25:30 crc kubenswrapper[4690]: I1203 14:25:30.381429 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8f2638db-cd46-439c-bf63-4cd912784601/glance-log/0.log" Dec 03 14:25:30 crc kubenswrapper[4690]: I1203 14:25:30.427927 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8f2638db-cd46-439c-bf63-4cd912784601/glance-httpd/0.log" Dec 03 14:25:30 crc kubenswrapper[4690]: I1203 14:25:30.487701 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-2n7xr_b529c361-6189-47c6-9c62-ea7fc0e8e30f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:30 crc kubenswrapper[4690]: I1203 14:25:30.631927 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kglnb_de573b60-ca9e-483c-8b4a-6d62fc6f56eb/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:30 crc kubenswrapper[4690]: I1203 14:25:30.925842 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412841-vn5nx_99ea4ce8-3695-4f3d-97fd-99d0e0b87029/keystone-cron/0.log" Dec 03 14:25:31 crc kubenswrapper[4690]: I1203 14:25:31.003434 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-788df6b698-7gkrp_cfe4fb6b-dd10-4462-9db0-dc2bc39f5dc4/keystone-api/0.log" Dec 03 14:25:31 crc kubenswrapper[4690]: I1203 14:25:31.152039 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9869e3c2-67c6-47be-8f00-45999ffcd3e7/kube-state-metrics/0.log" Dec 03 14:25:31 crc kubenswrapper[4690]: I1203 14:25:31.227113 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-8mcgq_8ce19539-f48f-446b-8a79-d528d58a9e8d/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:31 crc kubenswrapper[4690]: I1203 14:25:31.417739 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-tgvwl_f66f8700-77ac-4bed-b21b-8866f0edc58b/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:32 crc kubenswrapper[4690]: I1203 14:25:32.160225 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d87869f9-cmzfc_7d486acd-1716-4926-be13-3e67b3106c02/neutron-httpd/0.log" Dec 03 14:25:32 crc kubenswrapper[4690]: I1203 14:25:32.315741 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:25:32 crc kubenswrapper[4690]: E1203 14:25:32.316007 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:25:32 crc kubenswrapper[4690]: I1203 14:25:32.420730 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d87869f9-cmzfc_7d486acd-1716-4926-be13-3e67b3106c02/neutron-api/0.log" Dec 03 14:25:32 crc kubenswrapper[4690]: I1203 14:25:32.805879 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-2fszz_b1d5dc5c-d0a9-4194-88df-b48e39cad7d8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:33 crc kubenswrapper[4690]: I1203 14:25:33.436143 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278/nova-api-log/0.log" Dec 03 14:25:33 crc kubenswrapper[4690]: I1203 14:25:33.458142 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a730fe99-3893-4697-b946-f58470a93407/nova-cell0-conductor-conductor/0.log" Dec 03 14:25:33 crc kubenswrapper[4690]: I1203 14:25:33.782172 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9fdd9c06-8ef0-493a-b7f7-1cd4a4f93278/nova-api-api/0.log" Dec 03 14:25:34 crc kubenswrapper[4690]: I1203 14:25:34.323968 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_3141a41c-66ea-41fe-bc21-dd115b3cb98e/nova-cell1-conductor-conductor/0.log" Dec 03 14:25:34 crc kubenswrapper[4690]: I1203 14:25:34.354514 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d9db005b-de33-4de4-b7fb-c60783700345/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 14:25:34 crc kubenswrapper[4690]: I1203 14:25:34.474715 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hgcm8_6b06fe71-66c9-48a8-9e96-b9bdb23daf92/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:34 crc kubenswrapper[4690]: I1203 14:25:34.722745 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3f0b2df1-b5ed-498c-b6dd-784a32f97b3f/nova-metadata-log/0.log" Dec 03 14:25:35 crc kubenswrapper[4690]: I1203 14:25:35.008615 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c157eb34-b68f-4510-aafd-0c5e73435efe/mysql-bootstrap/0.log" Dec 03 14:25:35 crc kubenswrapper[4690]: I1203 14:25:35.144064 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_cc0fda9d-61ad-4a67-8c22-1da2e41949b0/nova-scheduler-scheduler/0.log" Dec 03 14:25:35 crc kubenswrapper[4690]: I1203 14:25:35.300415 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c157eb34-b68f-4510-aafd-0c5e73435efe/mysql-bootstrap/0.log" Dec 03 14:25:35 crc kubenswrapper[4690]: I1203 14:25:35.328579 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c157eb34-b68f-4510-aafd-0c5e73435efe/galera/0.log" Dec 03 14:25:35 crc kubenswrapper[4690]: I1203 14:25:35.563993 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5d47d61-0012-4eb8-ba88-da731d54ad45/mysql-bootstrap/0.log" Dec 03 14:25:36 crc kubenswrapper[4690]: I1203 14:25:36.090213 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5d47d61-0012-4eb8-ba88-da731d54ad45/mysql-bootstrap/0.log" Dec 03 14:25:36 crc kubenswrapper[4690]: I1203 14:25:36.183963 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5d47d61-0012-4eb8-ba88-da731d54ad45/galera/0.log" Dec 03 14:25:36 crc kubenswrapper[4690]: I1203 14:25:36.334237 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f47b845f-82f3-44b9-966d-eb3076f05362/openstackclient/0.log" Dec 03 14:25:36 crc kubenswrapper[4690]: I1203 14:25:36.362576 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3f0b2df1-b5ed-498c-b6dd-784a32f97b3f/nova-metadata-metadata/0.log" Dec 03 14:25:36 crc kubenswrapper[4690]: I1203 14:25:36.484533 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-wjlxx_338580e8-66c9-4169-93f5-1e7f530de31c/openstack-network-exporter/0.log" Dec 03 14:25:36 crc kubenswrapper[4690]: I1203 14:25:36.626551 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-mf8xw_64710eed-cdda-4f62-a007-766dd92f0b29/ovn-controller/0.log" Dec 03 14:25:36 crc kubenswrapper[4690]: I1203 14:25:36.799024 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-znnxj_cf553f8a-b61b-469b-9c0c-d602d73ef4c2/ovsdb-server-init/0.log" Dec 03 14:25:36 crc kubenswrapper[4690]: I1203 14:25:36.981763 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-znnxj_cf553f8a-b61b-469b-9c0c-d602d73ef4c2/ovsdb-server-init/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.095515 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-znnxj_cf553f8a-b61b-469b-9c0c-d602d73ef4c2/ovsdb-server/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.125039 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-znnxj_cf553f8a-b61b-469b-9c0c-d602d73ef4c2/ovs-vswitchd/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.326352 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-6pqsk_2ba8f301-845c-4425-9a12-72babfccd551/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.416343 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a172f1ad-a0fb-4f37-92e0-2c7267d11fa8/openstack-network-exporter/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.434560 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a172f1ad-a0fb-4f37-92e0-2c7267d11fa8/ovn-northd/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.630231 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c44ebdaa-211b-4ac6-90c6-5236eaee6295/openstack-network-exporter/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.640069 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c44ebdaa-211b-4ac6-90c6-5236eaee6295/ovsdbserver-nb/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.913921 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d19eec48-0864-43d8-8071-1aef21a26286/openstack-network-exporter/0.log" Dec 03 14:25:37 crc kubenswrapper[4690]: I1203 14:25:37.927522 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d19eec48-0864-43d8-8071-1aef21a26286/ovsdbserver-sb/0.log" Dec 03 14:25:38 crc kubenswrapper[4690]: I1203 14:25:38.159072 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f775c89b8-4tcrc_f9a732dd-555c-4e51-a334-80dab3104192/placement-api/0.log" Dec 03 14:25:38 crc kubenswrapper[4690]: I1203 14:25:38.252607 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8daef7dd-5b47-4f5c-8dc8-0da54d682e1d/setup-container/0.log" Dec 03 14:25:38 crc kubenswrapper[4690]: I1203 14:25:38.315088 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f775c89b8-4tcrc_f9a732dd-555c-4e51-a334-80dab3104192/placement-log/0.log" Dec 03 14:25:38 crc kubenswrapper[4690]: I1203 14:25:38.520922 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8daef7dd-5b47-4f5c-8dc8-0da54d682e1d/setup-container/0.log" Dec 03 14:25:38 crc kubenswrapper[4690]: I1203 14:25:38.593778 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8daef7dd-5b47-4f5c-8dc8-0da54d682e1d/rabbitmq/0.log" Dec 03 14:25:38 crc kubenswrapper[4690]: I1203 14:25:38.616922 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8/setup-container/0.log" Dec 03 14:25:39 crc kubenswrapper[4690]: I1203 14:25:39.142982 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7k2dt_0d10ad8c-5b1c-45c2-890a-eb42c648d437/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:39 crc kubenswrapper[4690]: I1203 14:25:39.159547 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8/setup-container/0.log" Dec 03 14:25:39 crc kubenswrapper[4690]: I1203 14:25:39.212349 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bbf43635-7ae1-4f01-b37f-7bdff8e5c4f8/rabbitmq/0.log" Dec 03 14:25:39 crc kubenswrapper[4690]: I1203 14:25:39.455212 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-f924x_d4625caf-d841-4682-ae69-e7d2d90ef220/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:39 crc kubenswrapper[4690]: I1203 14:25:39.608754 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-4mtxm_154bbd71-1585-49b0-84ab-3d11d49c99e6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:39 crc kubenswrapper[4690]: I1203 14:25:39.709243 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-2qlb4_80400476-b39e-4e4f-b57f-8813951d0f1a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:39 crc kubenswrapper[4690]: I1203 14:25:39.867993 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-gvv8v_91d255ec-88b0-416f-a990-92d930335b46/ssh-known-hosts-edpm-deployment/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.150654 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696775c5d7-q6gwf_9975bb19-d67e-4a25-a69f-8c1e37653961/proxy-server/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.243072 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696775c5d7-q6gwf_9975bb19-d67e-4a25-a69f-8c1e37653961/proxy-httpd/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.317586 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bl55c_262bf571-0151-4c11-a05c-1270656144f9/swift-ring-rebalance/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.477635 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/account-auditor/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.498257 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/account-reaper/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.639849 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/account-replicator/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.747209 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/container-auditor/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.772635 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/account-server/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.842611 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/container-replicator/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.944962 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/container-server/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.962684 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_187a8435-f796-4435-a0b0-211daf924823/memcached/0.log" Dec 03 14:25:40 crc kubenswrapper[4690]: I1203 14:25:40.984662 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/container-updater/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.023851 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-auditor/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.070136 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-expirer/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.190921 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-replicator/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.235932 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-updater/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.238410 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/object-server/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.279045 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/rsync/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.343605 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_28748086-659b-453c-bb91-8d9204374da6/swift-recon-cron/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.481844 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-m9q6n_c21f8ec4-b27d-4a89-bd74-e03efa0afe48/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.538448 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_46b67a35-a83b-48dd-a6ea-4464ff4ee3d5/tempest-tests-tempest-tests-runner/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.698010 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_7b33ef02-9b68-41ad-90ed-1adfe0cd8312/test-operator-logs-container/0.log" Dec 03 14:25:41 crc kubenswrapper[4690]: I1203 14:25:41.737564 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-l4bsr_96b45681-d5bf-474b-ac2c-2f8df92494d9/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 14:25:44 crc kubenswrapper[4690]: I1203 14:25:44.326016 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:25:44 crc kubenswrapper[4690]: E1203 14:25:44.326950 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:25:57 crc kubenswrapper[4690]: I1203 14:25:57.315435 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:25:57 crc kubenswrapper[4690]: E1203 14:25:57.316449 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:26:09 crc kubenswrapper[4690]: I1203 14:26:09.695625 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/util/0.log" Dec 03 14:26:09 crc kubenswrapper[4690]: I1203 14:26:09.905993 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/pull/0.log" Dec 03 14:26:09 crc kubenswrapper[4690]: I1203 14:26:09.906075 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/pull/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.058083 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/util/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.058356 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/util/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.095290 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/pull/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.119982 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02qjmkz_27e2f6d9-803d-48e7-ac24-bc3093175534/extract/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.242883 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-hx22v_1239bbbc-6613-4e1e-867b-fd62dee3c32b/kube-rbac-proxy/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.314667 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:26:10 crc kubenswrapper[4690]: E1203 14:26:10.314951 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.389594 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-hx22v_1239bbbc-6613-4e1e-867b-fd62dee3c32b/manager/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.437712 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pnhhq_bc776d95-7342-47e3-b2ad-abc4a3f7aa23/kube-rbac-proxy/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.519432 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pnhhq_bc776d95-7342-47e3-b2ad-abc4a3f7aa23/manager/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.600381 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-hwxjb_02c9e233-e41c-4fd2-9165-0ad5413fdaa2/kube-rbac-proxy/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.665449 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-hwxjb_02c9e233-e41c-4fd2-9165-0ad5413fdaa2/manager/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.823357 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-jmz4q_3b3afb5d-aef4-4c05-876f-52fa03532fe1/kube-rbac-proxy/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.885459 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-jmz4q_3b3afb5d-aef4-4c05-876f-52fa03532fe1/manager/0.log" Dec 03 14:26:10 crc kubenswrapper[4690]: I1203 14:26:10.983806 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-k6ht5_9da3821b-8edf-46c2-ac1c-5cabd4124968/kube-rbac-proxy/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.026764 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-k6ht5_9da3821b-8edf-46c2-ac1c-5cabd4124968/manager/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.187729 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gtcwg_4a193df2-0e41-48f0-a997-6f4b4e65bfe6/kube-rbac-proxy/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.194172 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gtcwg_4a193df2-0e41-48f0-a997-6f4b4e65bfe6/manager/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.334515 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-rrqhz_729e315d-2bf3-4568-941e-98ef7ad281eb/kube-rbac-proxy/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.447042 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-kshv8_5f4f6cc8-0d38-4c2a-b4f5-10a870c48479/kube-rbac-proxy/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.578549 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-rrqhz_729e315d-2bf3-4568-941e-98ef7ad281eb/manager/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.588085 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-kshv8_5f4f6cc8-0d38-4c2a-b4f5-10a870c48479/manager/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.654460 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-49f9d_0e42df6f-0109-4a0b-afeb-630bff5b8c35/kube-rbac-proxy/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.815652 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-49f9d_0e42df6f-0109-4a0b-afeb-630bff5b8c35/manager/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.910978 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-jk4h5_c731a6c0-1c65-45e1-893d-f620e83b5363/kube-rbac-proxy/0.log" Dec 03 14:26:11 crc kubenswrapper[4690]: I1203 14:26:11.991125 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-jk4h5_c731a6c0-1c65-45e1-893d-f620e83b5363/manager/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.049000 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-6g8xn_abb876f0-132b-4224-a3a7-074beba3367b/kube-rbac-proxy/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.119455 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-6g8xn_abb876f0-132b-4224-a3a7-074beba3367b/manager/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.304368 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-p559w_87cb4391-ed76-4f7a-a4b5-481d409bba7d/kube-rbac-proxy/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.346212 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-p559w_87cb4391-ed76-4f7a-a4b5-481d409bba7d/manager/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.535948 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5fll5_942df52e-3f08-47ef-bbeb-fe1191ced3e0/kube-rbac-proxy/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.627324 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5fll5_942df52e-3f08-47ef-bbeb-fe1191ced3e0/manager/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.697810 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-6q5ss_9890f67c-68ed-475b-9089-262d1fc270b4/kube-rbac-proxy/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.745783 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-6q5ss_9890f67c-68ed-475b-9089-262d1fc270b4/manager/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.856248 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd_f8b73cbf-3a07-46fa-98a4-db492ee1b711/kube-rbac-proxy/0.log" Dec 03 14:26:12 crc kubenswrapper[4690]: I1203 14:26:12.906823 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4dw8sd_f8b73cbf-3a07-46fa-98a4-db492ee1b711/manager/0.log" Dec 03 14:26:13 crc kubenswrapper[4690]: I1203 14:26:13.379307 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-58bc846945-vrqwj_b3182947-b66b-4777-8410-b6ffc37aa5d0/operator/0.log" Dec 03 14:26:13 crc kubenswrapper[4690]: I1203 14:26:13.485331 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-95zhf_ff4c17ee-a756-4743-aeb8-aae8fe520712/registry-server/0.log" Dec 03 14:26:13 crc kubenswrapper[4690]: I1203 14:26:13.709671 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lmq2n_7ef3f1c5-17d7-42f0-949b-1dbab5568051/kube-rbac-proxy/0.log" Dec 03 14:26:13 crc kubenswrapper[4690]: I1203 14:26:13.843161 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lmq2n_7ef3f1c5-17d7-42f0-949b-1dbab5568051/manager/0.log" Dec 03 14:26:13 crc kubenswrapper[4690]: I1203 14:26:13.976797 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7xsmn_55e64346-52e8-49b2-81de-057a98be5150/kube-rbac-proxy/0.log" Dec 03 14:26:13 crc kubenswrapper[4690]: I1203 14:26:13.989491 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7xsmn_55e64346-52e8-49b2-81de-057a98be5150/manager/0.log" Dec 03 14:26:14 crc kubenswrapper[4690]: I1203 14:26:14.509201 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7874b9874-9jdm5_953dae05-293d-488a-96d3-af4eda42ad47/manager/0.log" Dec 03 14:26:14 crc kubenswrapper[4690]: I1203 14:26:14.751529 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-fjthl_eb55cfd9-ad7f-473f-bbc6-dbb7bc436a58/operator/0.log" Dec 03 14:26:14 crc kubenswrapper[4690]: I1203 14:26:14.775825 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6xs6t_00aeb5e7-0973-4852-9fcf-6a51c3ef9720/kube-rbac-proxy/0.log" Dec 03 14:26:14 crc kubenswrapper[4690]: I1203 14:26:14.797563 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6xs6t_00aeb5e7-0973-4852-9fcf-6a51c3ef9720/manager/0.log" Dec 03 14:26:14 crc kubenswrapper[4690]: I1203 14:26:14.958606 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d8d6bf864-kzk67_7df25165-6b23-4dda-8105-05a933bc3ac8/kube-rbac-proxy/0.log" Dec 03 14:26:15 crc kubenswrapper[4690]: I1203 14:26:15.032984 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d8d6bf864-kzk67_7df25165-6b23-4dda-8105-05a933bc3ac8/manager/0.log" Dec 03 14:26:15 crc kubenswrapper[4690]: I1203 14:26:15.052767 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-x7xbr_0e2ce38d-747d-408f-8768-48a7310c0f9c/kube-rbac-proxy/0.log" Dec 03 14:26:15 crc kubenswrapper[4690]: I1203 14:26:15.149892 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-x7xbr_0e2ce38d-747d-408f-8768-48a7310c0f9c/manager/0.log" Dec 03 14:26:15 crc kubenswrapper[4690]: I1203 14:26:15.220500 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-skzwp_172f574f-c4bf-450a-8da9-fb4b879275f9/kube-rbac-proxy/0.log" Dec 03 14:26:15 crc kubenswrapper[4690]: I1203 14:26:15.258335 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-skzwp_172f574f-c4bf-450a-8da9-fb4b879275f9/manager/0.log" Dec 03 14:26:22 crc kubenswrapper[4690]: I1203 14:26:22.315368 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:26:22 crc kubenswrapper[4690]: E1203 14:26:22.316484 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:26:35 crc kubenswrapper[4690]: I1203 14:26:35.315148 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:26:35 crc kubenswrapper[4690]: E1203 14:26:35.316080 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:26:36 crc kubenswrapper[4690]: I1203 14:26:36.424323 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jvcdk_2fb0cc5a-0d3d-47ca-8ccf-c4eee70f01b3/control-plane-machine-set-operator/0.log" Dec 03 14:26:36 crc kubenswrapper[4690]: I1203 14:26:36.598314 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jldld_fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26/kube-rbac-proxy/0.log" Dec 03 14:26:36 crc kubenswrapper[4690]: I1203 14:26:36.628545 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jldld_fe68d0d8-83c1-4cd7-b82b-ae11a3eeff26/machine-api-operator/0.log" Dec 03 14:26:47 crc kubenswrapper[4690]: I1203 14:26:47.315129 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:26:47 crc kubenswrapper[4690]: E1203 14:26:47.315982 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:26:50 crc kubenswrapper[4690]: I1203 14:26:50.787585 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2gp2s_9c22c662-9a32-4721-9958-12651fa443a7/cert-manager-controller/0.log" Dec 03 14:26:50 crc kubenswrapper[4690]: I1203 14:26:50.798682 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-2p8k2_6daa0e47-6f3f-4ac8-b1a8-596401c33ef2/cert-manager-cainjector/0.log" Dec 03 14:26:50 crc kubenswrapper[4690]: I1203 14:26:50.992848 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4jfnl_f47a381d-4adf-4543-b4f3-793cdc3db649/cert-manager-webhook/0.log" Dec 03 14:27:00 crc kubenswrapper[4690]: I1203 14:27:00.315761 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:27:00 crc kubenswrapper[4690]: E1203 14:27:00.316681 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:27:03 crc kubenswrapper[4690]: I1203 14:27:03.627608 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-4xg2d_bb3da90d-da37-49ad-ad46-230759602e46/nmstate-console-plugin/0.log" Dec 03 14:27:03 crc kubenswrapper[4690]: I1203 14:27:03.777833 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-v429n_89d8054c-b35f-42fd-960d-16e1b39ad983/nmstate-handler/0.log" Dec 03 14:27:03 crc kubenswrapper[4690]: I1203 14:27:03.826561 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-hjrl4_33310d90-a18c-44db-a7f3-37f2086de63a/kube-rbac-proxy/0.log" Dec 03 14:27:03 crc kubenswrapper[4690]: I1203 14:27:03.887067 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-hjrl4_33310d90-a18c-44db-a7f3-37f2086de63a/nmstate-metrics/0.log" Dec 03 14:27:04 crc kubenswrapper[4690]: I1203 14:27:04.025225 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-p69d4_0aacb971-67b1-4e6a-a89c-7aa8b404f046/nmstate-operator/0.log" Dec 03 14:27:04 crc kubenswrapper[4690]: I1203 14:27:04.134002 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-mw7tn_6020072e-2f65-4d13-85be-d300b4ec7f6d/nmstate-webhook/0.log" Dec 03 14:27:13 crc kubenswrapper[4690]: I1203 14:27:13.315577 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:27:13 crc kubenswrapper[4690]: E1203 14:27:13.316656 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:27:17 crc kubenswrapper[4690]: I1203 14:27:17.450779 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-bb5b4b448-ql656_72a896a4-41ed-4652-bfa5-358369eddfa2/kube-rbac-proxy/0.log" Dec 03 14:27:17 crc kubenswrapper[4690]: I1203 14:27:17.571582 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-bb5b4b448-ql656_72a896a4-41ed-4652-bfa5-358369eddfa2/manager/0.log" Dec 03 14:27:24 crc kubenswrapper[4690]: I1203 14:27:24.323238 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:27:24 crc kubenswrapper[4690]: E1203 14:27:24.324112 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:27:33 crc kubenswrapper[4690]: I1203 14:27:33.497267 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-lqf4z_fb39061a-2c4c-4c29-95b6-6f070367df86/cluster-logging-operator/0.log" Dec 03 14:27:33 crc kubenswrapper[4690]: I1203 14:27:33.697947 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-g89tt_ed87f330-d809-4183-8cc7-61caf74208a8/collector/0.log" Dec 03 14:27:33 crc kubenswrapper[4690]: I1203 14:27:33.762780 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_6289c341-366c-4b2e-a4ed-2d0eb43d0da4/loki-compactor/0.log" Dec 03 14:27:33 crc kubenswrapper[4690]: I1203 14:27:33.878669 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-qx2hh_d4eb66c8-ca39-4ba9-8717-a488817c5304/loki-distributor/0.log" Dec 03 14:27:33 crc kubenswrapper[4690]: I1203 14:27:33.926768 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-79d97f9547-8tgpn_4b3976e3-672e-4117-a9c6-9a7f34d42be7/gateway/0.log" Dec 03 14:27:33 crc kubenswrapper[4690]: I1203 14:27:33.979714 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-79d97f9547-8tgpn_4b3976e3-672e-4117-a9c6-9a7f34d42be7/opa/0.log" Dec 03 14:27:34 crc kubenswrapper[4690]: I1203 14:27:34.118285 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-79d97f9547-sps84_f1541c44-fb2e-4414-88f0-21e464b629f0/gateway/0.log" Dec 03 14:27:34 crc kubenswrapper[4690]: I1203 14:27:34.169932 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-79d97f9547-sps84_f1541c44-fb2e-4414-88f0-21e464b629f0/opa/0.log" Dec 03 14:27:34 crc kubenswrapper[4690]: I1203 14:27:34.315071 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_e56039ef-9919-4d3d-9b04-876bad0f5756/loki-index-gateway/0.log" Dec 03 14:27:34 crc kubenswrapper[4690]: I1203 14:27:34.847557 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-ckhv9_36825bd7-9921-4f76-a26d-6dd05f00bc3e/loki-querier/0.log" Dec 03 14:27:34 crc kubenswrapper[4690]: I1203 14:27:34.895352 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_ecc469a1-b69c-4cbd-bc8f-c85e5f68ee86/loki-ingester/0.log" Dec 03 14:27:35 crc kubenswrapper[4690]: I1203 14:27:35.043527 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-pslms_1b8cb794-f99d-41b0-bbdc-ed052be8f4a5/loki-query-frontend/0.log" Dec 03 14:27:35 crc kubenswrapper[4690]: I1203 14:27:35.315000 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:27:35 crc kubenswrapper[4690]: E1203 14:27:35.315402 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:27:47 crc kubenswrapper[4690]: I1203 14:27:47.315387 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:27:47 crc kubenswrapper[4690]: E1203 14:27:47.316590 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.173799 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t4xcd_e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1/kube-rbac-proxy/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.310709 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t4xcd_e65a2f57-73b0-4eb7-a62e-c8f4b8f816a1/controller/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.409728 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-frr-files/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.594961 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-frr-files/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.636896 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-metrics/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.661344 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-reloader/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.675021 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-reloader/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.851399 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-frr-files/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.890204 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-reloader/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.915014 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-metrics/0.log" Dec 03 14:27:49 crc kubenswrapper[4690]: I1203 14:27:49.939975 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-metrics/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.083305 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-reloader/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.094088 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-metrics/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.117403 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/cp-frr-files/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.206044 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/controller/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.296223 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/frr-metrics/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.342254 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/kube-rbac-proxy/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.470522 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/kube-rbac-proxy-frr/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.531173 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/reloader/0.log" Dec 03 14:27:50 crc kubenswrapper[4690]: I1203 14:27:50.795301 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-r8zcv_ece18280-b324-4c1e-b3d5-411e5c146869/frr-k8s-webhook-server/0.log" Dec 03 14:27:51 crc kubenswrapper[4690]: I1203 14:27:51.006105 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c954657ff-b5h2m_93d03eb3-9a1d-452e-8812-c5a871652ab3/webhook-server/0.log" Dec 03 14:27:51 crc kubenswrapper[4690]: I1203 14:27:51.013929 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5994f497d6-5c9m6_6b62fccc-0740-4e1f-8efb-61bac3a9261e/manager/0.log" Dec 03 14:27:51 crc kubenswrapper[4690]: I1203 14:27:51.351204 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7hmjp_e1b3ec2d-089c-409a-901e-4f466ec07855/kube-rbac-proxy/0.log" Dec 03 14:27:51 crc kubenswrapper[4690]: I1203 14:27:51.858539 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hfmdl_efd8bd0a-d3e9-44ea-a1d7-aab093dea9f2/frr/0.log" Dec 03 14:27:51 crc kubenswrapper[4690]: I1203 14:27:51.911361 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7hmjp_e1b3ec2d-089c-409a-901e-4f466ec07855/speaker/0.log" Dec 03 14:28:00 crc kubenswrapper[4690]: I1203 14:28:00.316246 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:28:00 crc kubenswrapper[4690]: E1203 14:28:00.317207 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:28:06 crc kubenswrapper[4690]: I1203 14:28:06.108305 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/util/0.log" Dec 03 14:28:06 crc kubenswrapper[4690]: I1203 14:28:06.370935 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/pull/0.log" Dec 03 14:28:06 crc kubenswrapper[4690]: I1203 14:28:06.378663 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/util/0.log" Dec 03 14:28:06 crc kubenswrapper[4690]: I1203 14:28:06.407899 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/pull/0.log" Dec 03 14:28:06 crc kubenswrapper[4690]: I1203 14:28:06.644697 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/util/0.log" Dec 03 14:28:06 crc kubenswrapper[4690]: I1203 14:28:06.692962 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/extract/0.log" Dec 03 14:28:06 crc kubenswrapper[4690]: I1203 14:28:06.751990 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb876dwx_d67d533a-906d-49f6-8331-7bc5f85bff55/pull/0.log" Dec 03 14:28:06 crc kubenswrapper[4690]: I1203 14:28:06.869835 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/util/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.094816 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/pull/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.104035 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/util/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.128536 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/pull/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.563667 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/extract/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.687212 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/util/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.689807 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmjpbl_314999b8-63d0-4f07-8b80-2aa04364cb5a/pull/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.753581 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/util/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.982291 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/pull/0.log" Dec 03 14:28:07 crc kubenswrapper[4690]: I1203 14:28:07.992880 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/util/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.005604 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/pull/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.188098 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/pull/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.211908 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/util/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.276511 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f2j8sx_2d467dd5-228e-436a-8fa0-c96bbf0ce110/extract/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.420521 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/util/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.576597 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/util/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.603823 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/pull/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.611275 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/pull/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.859105 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/pull/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.894208 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/extract/0.log" Dec 03 14:28:08 crc kubenswrapper[4690]: I1203 14:28:08.898005 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cjqng_64127212-43bb-457c-b18d-3be4ea85d178/util/0.log" Dec 03 14:28:09 crc kubenswrapper[4690]: I1203 14:28:09.114789 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-utilities/0.log" Dec 03 14:28:09 crc kubenswrapper[4690]: I1203 14:28:09.358999 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-content/0.log" Dec 03 14:28:09 crc kubenswrapper[4690]: I1203 14:28:09.382338 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-content/0.log" Dec 03 14:28:09 crc kubenswrapper[4690]: I1203 14:28:09.401222 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-utilities/0.log" Dec 03 14:28:09 crc kubenswrapper[4690]: I1203 14:28:09.582753 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-utilities/0.log" Dec 03 14:28:09 crc kubenswrapper[4690]: I1203 14:28:09.629828 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/extract-content/0.log" Dec 03 14:28:09 crc kubenswrapper[4690]: I1203 14:28:09.852054 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-utilities/0.log" Dec 03 14:28:10 crc kubenswrapper[4690]: I1203 14:28:10.158041 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-content/0.log" Dec 03 14:28:10 crc kubenswrapper[4690]: I1203 14:28:10.233960 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-content/0.log" Dec 03 14:28:10 crc kubenswrapper[4690]: I1203 14:28:10.243603 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-utilities/0.log" Dec 03 14:28:10 crc kubenswrapper[4690]: I1203 14:28:10.323729 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bbjwm_4a384d66-783b-45bc-afa5-fb9d77e223ac/registry-server/0.log" Dec 03 14:28:10 crc kubenswrapper[4690]: I1203 14:28:10.493657 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-content/0.log" Dec 03 14:28:10 crc kubenswrapper[4690]: I1203 14:28:10.507541 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/extract-utilities/0.log" Dec 03 14:28:10 crc kubenswrapper[4690]: I1203 14:28:10.642565 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-t8hvp_2b3e95ae-d895-45bf-9abd-73cf60e5f139/marketplace-operator/0.log" Dec 03 14:28:10 crc kubenswrapper[4690]: I1203 14:28:10.802665 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-utilities/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.261100 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-content/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.272983 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z9fw6_096d833c-6989-4a0e-8437-bdbc751a603d/registry-server/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.328172 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-utilities/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.411065 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-content/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.532274 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-utilities/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.598238 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/extract-content/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.711441 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-utilities/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.794098 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w9wd6_d6a489f2-fe20-4d26-87f4-82c65ae7a461/registry-server/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.977154 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-utilities/0.log" Dec 03 14:28:11 crc kubenswrapper[4690]: I1203 14:28:11.988584 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-content/0.log" Dec 03 14:28:12 crc kubenswrapper[4690]: I1203 14:28:12.020523 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-content/0.log" Dec 03 14:28:12 crc kubenswrapper[4690]: I1203 14:28:12.234058 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-utilities/0.log" Dec 03 14:28:12 crc kubenswrapper[4690]: I1203 14:28:12.242450 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/extract-content/0.log" Dec 03 14:28:12 crc kubenswrapper[4690]: I1203 14:28:12.315247 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:28:12 crc kubenswrapper[4690]: E1203 14:28:12.316055 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:28:12 crc kubenswrapper[4690]: I1203 14:28:12.521805 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lgw8x_bd39d84c-a392-407c-872b-9be544eb1816/registry-server/0.log" Dec 03 14:28:23 crc kubenswrapper[4690]: I1203 14:28:23.315417 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:28:23 crc kubenswrapper[4690]: E1203 14:28:23.316719 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.169993 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jwjzn"] Dec 03 14:28:29 crc kubenswrapper[4690]: E1203 14:28:29.176230 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446b00f8-b816-4daa-903e-7f370be05883" containerName="container-00" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.176268 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="446b00f8-b816-4daa-903e-7f370be05883" containerName="container-00" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.176784 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="446b00f8-b816-4daa-903e-7f370be05883" containerName="container-00" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.178901 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.192456 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jwjzn"] Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.334336 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jt2x\" (UniqueName: \"kubernetes.io/projected/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-kube-api-access-7jt2x\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.334490 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-catalog-content\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.334601 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-utilities\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.436932 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jt2x\" (UniqueName: \"kubernetes.io/projected/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-kube-api-access-7jt2x\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.437105 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-catalog-content\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.437263 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-utilities\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.439342 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-catalog-content\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.440194 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-utilities\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.462835 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jt2x\" (UniqueName: \"kubernetes.io/projected/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-kube-api-access-7jt2x\") pod \"community-operators-jwjzn\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:29 crc kubenswrapper[4690]: I1203 14:28:29.522299 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:30 crc kubenswrapper[4690]: I1203 14:28:30.196926 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jwjzn"] Dec 03 14:28:31 crc kubenswrapper[4690]: I1203 14:28:31.024784 4690 generic.go:334] "Generic (PLEG): container finished" podID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerID="59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2" exitCode=0 Dec 03 14:28:31 crc kubenswrapper[4690]: I1203 14:28:31.025256 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwjzn" event={"ID":"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19","Type":"ContainerDied","Data":"59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2"} Dec 03 14:28:31 crc kubenswrapper[4690]: I1203 14:28:31.025295 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwjzn" event={"ID":"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19","Type":"ContainerStarted","Data":"efd99ff0a80d14aae78f43b8db33cbf89066633e1f67f3d3447ebc379ced1028"} Dec 03 14:28:31 crc kubenswrapper[4690]: I1203 14:28:31.027931 4690 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:28:32 crc kubenswrapper[4690]: I1203 14:28:32.039740 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwjzn" event={"ID":"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19","Type":"ContainerStarted","Data":"f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771"} Dec 03 14:28:33 crc kubenswrapper[4690]: I1203 14:28:33.058361 4690 generic.go:334] "Generic (PLEG): container finished" podID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerID="f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771" exitCode=0 Dec 03 14:28:33 crc kubenswrapper[4690]: I1203 14:28:33.058479 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwjzn" event={"ID":"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19","Type":"ContainerDied","Data":"f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771"} Dec 03 14:28:34 crc kubenswrapper[4690]: I1203 14:28:34.077053 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwjzn" event={"ID":"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19","Type":"ContainerStarted","Data":"b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96"} Dec 03 14:28:34 crc kubenswrapper[4690]: I1203 14:28:34.112670 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jwjzn" podStartSLOduration=2.6527042769999998 podStartE2EDuration="5.112643547s" podCreationTimestamp="2025-12-03 14:28:29 +0000 UTC" firstStartedPulling="2025-12-03 14:28:31.027579189 +0000 UTC m=+4757.008499622" lastFinishedPulling="2025-12-03 14:28:33.487518459 +0000 UTC m=+4759.468438892" observedRunningTime="2025-12-03 14:28:34.104383517 +0000 UTC m=+4760.085303950" watchObservedRunningTime="2025-12-03 14:28:34.112643547 +0000 UTC m=+4760.093563980" Dec 03 14:28:37 crc kubenswrapper[4690]: I1203 14:28:37.314969 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:28:37 crc kubenswrapper[4690]: E1203 14:28:37.315822 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:28:39 crc kubenswrapper[4690]: I1203 14:28:39.529275 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:39 crc kubenswrapper[4690]: I1203 14:28:39.530163 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:39 crc kubenswrapper[4690]: I1203 14:28:39.594061 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:40 crc kubenswrapper[4690]: I1203 14:28:40.196830 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:40 crc kubenswrapper[4690]: I1203 14:28:40.265274 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jwjzn"] Dec 03 14:28:42 crc kubenswrapper[4690]: I1203 14:28:42.163446 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jwjzn" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerName="registry-server" containerID="cri-o://b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96" gracePeriod=2 Dec 03 14:28:42 crc kubenswrapper[4690]: I1203 14:28:42.808083 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.010579 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jt2x\" (UniqueName: \"kubernetes.io/projected/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-kube-api-access-7jt2x\") pod \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.010943 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-catalog-content\") pod \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.011054 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-utilities\") pod \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\" (UID: \"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19\") " Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.012118 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-utilities" (OuterVolumeSpecName: "utilities") pod "8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" (UID: "8e1dc4ea-2bf2-41c2-a633-89e7c0271d19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.018045 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-kube-api-access-7jt2x" (OuterVolumeSpecName: "kube-api-access-7jt2x") pod "8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" (UID: "8e1dc4ea-2bf2-41c2-a633-89e7c0271d19"). InnerVolumeSpecName "kube-api-access-7jt2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.067094 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" (UID: "8e1dc4ea-2bf2-41c2-a633-89e7c0271d19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.114427 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.114465 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.114476 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jt2x\" (UniqueName: \"kubernetes.io/projected/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19-kube-api-access-7jt2x\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.179022 4690 generic.go:334] "Generic (PLEG): container finished" podID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerID="b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96" exitCode=0 Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.179085 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwjzn" event={"ID":"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19","Type":"ContainerDied","Data":"b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96"} Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.179102 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwjzn" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.179128 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwjzn" event={"ID":"8e1dc4ea-2bf2-41c2-a633-89e7c0271d19","Type":"ContainerDied","Data":"efd99ff0a80d14aae78f43b8db33cbf89066633e1f67f3d3447ebc379ced1028"} Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.179168 4690 scope.go:117] "RemoveContainer" containerID="b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.216539 4690 scope.go:117] "RemoveContainer" containerID="f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.224605 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jwjzn"] Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.236520 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jwjzn"] Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.241938 4690 scope.go:117] "RemoveContainer" containerID="59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.346489 4690 scope.go:117] "RemoveContainer" containerID="b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96" Dec 03 14:28:43 crc kubenswrapper[4690]: E1203 14:28:43.352589 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96\": container with ID starting with b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96 not found: ID does not exist" containerID="b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.352670 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96"} err="failed to get container status \"b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96\": rpc error: code = NotFound desc = could not find container \"b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96\": container with ID starting with b2f5f8767355d69265fe75c9c606d4a08d2f3649f3065c41df7ef59586afed96 not found: ID does not exist" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.352722 4690 scope.go:117] "RemoveContainer" containerID="f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771" Dec 03 14:28:43 crc kubenswrapper[4690]: E1203 14:28:43.353592 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771\": container with ID starting with f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771 not found: ID does not exist" containerID="f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.353655 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771"} err="failed to get container status \"f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771\": rpc error: code = NotFound desc = could not find container \"f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771\": container with ID starting with f15ab1fe6bb66dd1782d8b1f35a3b868190adfab014482b723e7f4b4b3441771 not found: ID does not exist" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.353697 4690 scope.go:117] "RemoveContainer" containerID="59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2" Dec 03 14:28:43 crc kubenswrapper[4690]: E1203 14:28:43.354321 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2\": container with ID starting with 59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2 not found: ID does not exist" containerID="59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2" Dec 03 14:28:43 crc kubenswrapper[4690]: I1203 14:28:43.354356 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2"} err="failed to get container status \"59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2\": rpc error: code = NotFound desc = could not find container \"59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2\": container with ID starting with 59722d8a89719c089a2b539c01ac0ef88330c01e6f24e385bf34d3845ecb99d2 not found: ID does not exist" Dec 03 14:28:44 crc kubenswrapper[4690]: I1203 14:28:44.329694 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" path="/var/lib/kubelet/pods/8e1dc4ea-2bf2-41c2-a633-89e7c0271d19/volumes" Dec 03 14:28:45 crc kubenswrapper[4690]: I1203 14:28:45.489786 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-bb5b4b448-ql656_72a896a4-41ed-4652-bfa5-358369eddfa2/kube-rbac-proxy/0.log" Dec 03 14:28:45 crc kubenswrapper[4690]: I1203 14:28:45.597501 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-bb5b4b448-ql656_72a896a4-41ed-4652-bfa5-358369eddfa2/manager/0.log" Dec 03 14:28:51 crc kubenswrapper[4690]: I1203 14:28:51.316952 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:28:51 crc kubenswrapper[4690]: E1203 14:28:51.317947 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:29:05 crc kubenswrapper[4690]: I1203 14:29:05.315282 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:29:05 crc kubenswrapper[4690]: E1203 14:29:05.316341 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:29:18 crc kubenswrapper[4690]: I1203 14:29:18.315786 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:29:18 crc kubenswrapper[4690]: E1203 14:29:18.316685 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.250290 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p4qqg"] Dec 03 14:29:20 crc kubenswrapper[4690]: E1203 14:29:20.252281 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerName="extract-utilities" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.252371 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerName="extract-utilities" Dec 03 14:29:20 crc kubenswrapper[4690]: E1203 14:29:20.252505 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerName="extract-content" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.252572 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerName="extract-content" Dec 03 14:29:20 crc kubenswrapper[4690]: E1203 14:29:20.252649 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerName="registry-server" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.252707 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerName="registry-server" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.253068 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e1dc4ea-2bf2-41c2-a633-89e7c0271d19" containerName="registry-server" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.254971 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.265795 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p4qqg"] Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.360734 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmgz5\" (UniqueName: \"kubernetes.io/projected/4ff2a347-dcea-4986-8fc6-a2599e74307d-kube-api-access-qmgz5\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.361364 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-utilities\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.361722 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-catalog-content\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.464560 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmgz5\" (UniqueName: \"kubernetes.io/projected/4ff2a347-dcea-4986-8fc6-a2599e74307d-kube-api-access-qmgz5\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.464960 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-utilities\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.465134 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-catalog-content\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.465489 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-utilities\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.465536 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-catalog-content\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.855961 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmgz5\" (UniqueName: \"kubernetes.io/projected/4ff2a347-dcea-4986-8fc6-a2599e74307d-kube-api-access-qmgz5\") pod \"certified-operators-p4qqg\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:20 crc kubenswrapper[4690]: I1203 14:29:20.893594 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:21 crc kubenswrapper[4690]: I1203 14:29:21.449830 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p4qqg"] Dec 03 14:29:21 crc kubenswrapper[4690]: I1203 14:29:21.641884 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4qqg" event={"ID":"4ff2a347-dcea-4986-8fc6-a2599e74307d","Type":"ContainerStarted","Data":"4224d89f9bfcea4579723e3e6238b9d9c29275659d0d63a0c8b06acb48b930ec"} Dec 03 14:29:22 crc kubenswrapper[4690]: I1203 14:29:22.657896 4690 generic.go:334] "Generic (PLEG): container finished" podID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerID="d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e" exitCode=0 Dec 03 14:29:22 crc kubenswrapper[4690]: I1203 14:29:22.658021 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4qqg" event={"ID":"4ff2a347-dcea-4986-8fc6-a2599e74307d","Type":"ContainerDied","Data":"d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e"} Dec 03 14:29:24 crc kubenswrapper[4690]: I1203 14:29:24.683419 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4qqg" event={"ID":"4ff2a347-dcea-4986-8fc6-a2599e74307d","Type":"ContainerStarted","Data":"c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539"} Dec 03 14:29:26 crc kubenswrapper[4690]: I1203 14:29:26.716406 4690 generic.go:334] "Generic (PLEG): container finished" podID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerID="c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539" exitCode=0 Dec 03 14:29:26 crc kubenswrapper[4690]: I1203 14:29:26.716809 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4qqg" event={"ID":"4ff2a347-dcea-4986-8fc6-a2599e74307d","Type":"ContainerDied","Data":"c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539"} Dec 03 14:29:29 crc kubenswrapper[4690]: I1203 14:29:29.775159 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4qqg" event={"ID":"4ff2a347-dcea-4986-8fc6-a2599e74307d","Type":"ContainerStarted","Data":"5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2"} Dec 03 14:29:29 crc kubenswrapper[4690]: I1203 14:29:29.808546 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p4qqg" podStartSLOduration=4.045121218 podStartE2EDuration="9.808510062s" podCreationTimestamp="2025-12-03 14:29:20 +0000 UTC" firstStartedPulling="2025-12-03 14:29:22.660448021 +0000 UTC m=+4808.641368454" lastFinishedPulling="2025-12-03 14:29:28.423836855 +0000 UTC m=+4814.404757298" observedRunningTime="2025-12-03 14:29:29.800852217 +0000 UTC m=+4815.781772660" watchObservedRunningTime="2025-12-03 14:29:29.808510062 +0000 UTC m=+4815.789430495" Dec 03 14:29:30 crc kubenswrapper[4690]: I1203 14:29:30.895412 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:30 crc kubenswrapper[4690]: I1203 14:29:30.895841 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:30 crc kubenswrapper[4690]: I1203 14:29:30.968231 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:31 crc kubenswrapper[4690]: I1203 14:29:31.316530 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:29:31 crc kubenswrapper[4690]: E1203 14:29:31.316828 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:29:40 crc kubenswrapper[4690]: I1203 14:29:40.949614 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:41 crc kubenswrapper[4690]: I1203 14:29:41.012636 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p4qqg"] Dec 03 14:29:41 crc kubenswrapper[4690]: I1203 14:29:41.916313 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p4qqg" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerName="registry-server" containerID="cri-o://5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2" gracePeriod=2 Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.479672 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.539824 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-catalog-content\") pod \"4ff2a347-dcea-4986-8fc6-a2599e74307d\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.539902 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-utilities\") pod \"4ff2a347-dcea-4986-8fc6-a2599e74307d\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.540056 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmgz5\" (UniqueName: \"kubernetes.io/projected/4ff2a347-dcea-4986-8fc6-a2599e74307d-kube-api-access-qmgz5\") pod \"4ff2a347-dcea-4986-8fc6-a2599e74307d\" (UID: \"4ff2a347-dcea-4986-8fc6-a2599e74307d\") " Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.541613 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-utilities" (OuterVolumeSpecName: "utilities") pod "4ff2a347-dcea-4986-8fc6-a2599e74307d" (UID: "4ff2a347-dcea-4986-8fc6-a2599e74307d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.546657 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff2a347-dcea-4986-8fc6-a2599e74307d-kube-api-access-qmgz5" (OuterVolumeSpecName: "kube-api-access-qmgz5") pod "4ff2a347-dcea-4986-8fc6-a2599e74307d" (UID: "4ff2a347-dcea-4986-8fc6-a2599e74307d"). InnerVolumeSpecName "kube-api-access-qmgz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.598037 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ff2a347-dcea-4986-8fc6-a2599e74307d" (UID: "4ff2a347-dcea-4986-8fc6-a2599e74307d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.643190 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmgz5\" (UniqueName: \"kubernetes.io/projected/4ff2a347-dcea-4986-8fc6-a2599e74307d-kube-api-access-qmgz5\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.643476 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.643534 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff2a347-dcea-4986-8fc6-a2599e74307d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.932097 4690 generic.go:334] "Generic (PLEG): container finished" podID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerID="5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2" exitCode=0 Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.932162 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4qqg" event={"ID":"4ff2a347-dcea-4986-8fc6-a2599e74307d","Type":"ContainerDied","Data":"5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2"} Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.932204 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4qqg" event={"ID":"4ff2a347-dcea-4986-8fc6-a2599e74307d","Type":"ContainerDied","Data":"4224d89f9bfcea4579723e3e6238b9d9c29275659d0d63a0c8b06acb48b930ec"} Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.932224 4690 scope.go:117] "RemoveContainer" containerID="5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.932422 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4qqg" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.983954 4690 scope.go:117] "RemoveContainer" containerID="c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539" Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.985962 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p4qqg"] Dec 03 14:29:42 crc kubenswrapper[4690]: I1203 14:29:42.999590 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p4qqg"] Dec 03 14:29:43 crc kubenswrapper[4690]: I1203 14:29:43.014138 4690 scope.go:117] "RemoveContainer" containerID="d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e" Dec 03 14:29:43 crc kubenswrapper[4690]: I1203 14:29:43.062846 4690 scope.go:117] "RemoveContainer" containerID="5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2" Dec 03 14:29:43 crc kubenswrapper[4690]: E1203 14:29:43.063522 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2\": container with ID starting with 5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2 not found: ID does not exist" containerID="5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2" Dec 03 14:29:43 crc kubenswrapper[4690]: I1203 14:29:43.063574 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2"} err="failed to get container status \"5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2\": rpc error: code = NotFound desc = could not find container \"5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2\": container with ID starting with 5a2dee971418bb54eb717c46023ee5822f24b5c7e2e9f939aa2a7dfaa19848a2 not found: ID does not exist" Dec 03 14:29:43 crc kubenswrapper[4690]: I1203 14:29:43.063605 4690 scope.go:117] "RemoveContainer" containerID="c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539" Dec 03 14:29:43 crc kubenswrapper[4690]: E1203 14:29:43.064090 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539\": container with ID starting with c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539 not found: ID does not exist" containerID="c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539" Dec 03 14:29:43 crc kubenswrapper[4690]: I1203 14:29:43.064193 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539"} err="failed to get container status \"c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539\": rpc error: code = NotFound desc = could not find container \"c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539\": container with ID starting with c8d758caa7b5f952fafd7d73ad2ac4511eb4161ac095e0185c5b0f2aa65f3539 not found: ID does not exist" Dec 03 14:29:43 crc kubenswrapper[4690]: I1203 14:29:43.064273 4690 scope.go:117] "RemoveContainer" containerID="d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e" Dec 03 14:29:43 crc kubenswrapper[4690]: E1203 14:29:43.064641 4690 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e\": container with ID starting with d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e not found: ID does not exist" containerID="d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e" Dec 03 14:29:43 crc kubenswrapper[4690]: I1203 14:29:43.064729 4690 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e"} err="failed to get container status \"d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e\": rpc error: code = NotFound desc = could not find container \"d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e\": container with ID starting with d7fadeae513c1316174100485d1ccb4e4153da78d64707008ca829a87683b77e not found: ID does not exist" Dec 03 14:29:44 crc kubenswrapper[4690]: I1203 14:29:44.331153 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" path="/var/lib/kubelet/pods/4ff2a347-dcea-4986-8fc6-a2599e74307d/volumes" Dec 03 14:29:45 crc kubenswrapper[4690]: I1203 14:29:45.318030 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:29:45 crc kubenswrapper[4690]: E1203 14:29:45.319199 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:29:59 crc kubenswrapper[4690]: I1203 14:29:59.315538 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:29:59 crc kubenswrapper[4690]: E1203 14:29:59.316477 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.148037 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk"] Dec 03 14:30:00 crc kubenswrapper[4690]: E1203 14:30:00.149126 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerName="extract-content" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.149159 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerName="extract-content" Dec 03 14:30:00 crc kubenswrapper[4690]: E1203 14:30:00.149194 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerName="extract-utilities" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.149204 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerName="extract-utilities" Dec 03 14:30:00 crc kubenswrapper[4690]: E1203 14:30:00.149219 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerName="registry-server" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.149225 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerName="registry-server" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.149470 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ff2a347-dcea-4986-8fc6-a2599e74307d" containerName="registry-server" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.150327 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.153835 4690 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.154157 4690 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.169349 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk"] Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.276592 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq5sb\" (UniqueName: \"kubernetes.io/projected/41815399-ff92-43f3-af2d-3433fa3e73c9-kube-api-access-xq5sb\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.276682 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41815399-ff92-43f3-af2d-3433fa3e73c9-config-volume\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.276799 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41815399-ff92-43f3-af2d-3433fa3e73c9-secret-volume\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.379656 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41815399-ff92-43f3-af2d-3433fa3e73c9-secret-volume\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.379927 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq5sb\" (UniqueName: \"kubernetes.io/projected/41815399-ff92-43f3-af2d-3433fa3e73c9-kube-api-access-xq5sb\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.379970 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41815399-ff92-43f3-af2d-3433fa3e73c9-config-volume\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.381277 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41815399-ff92-43f3-af2d-3433fa3e73c9-config-volume\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.388108 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41815399-ff92-43f3-af2d-3433fa3e73c9-secret-volume\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.415270 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq5sb\" (UniqueName: \"kubernetes.io/projected/41815399-ff92-43f3-af2d-3433fa3e73c9-kube-api-access-xq5sb\") pod \"collect-profiles-29412870-lzvvk\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.478449 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:00 crc kubenswrapper[4690]: I1203 14:30:00.982648 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk"] Dec 03 14:30:01 crc kubenswrapper[4690]: I1203 14:30:01.130727 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" event={"ID":"41815399-ff92-43f3-af2d-3433fa3e73c9","Type":"ContainerStarted","Data":"26a0d31cb1c216376f3b542ddb03a6f75e48bf5202321318f3b067c2ef21c4f9"} Dec 03 14:30:02 crc kubenswrapper[4690]: I1203 14:30:02.154327 4690 generic.go:334] "Generic (PLEG): container finished" podID="41815399-ff92-43f3-af2d-3433fa3e73c9" containerID="b9a6f484acadba9194e493665e9bbe4adb44ed90bfbdec8fc7c95feb38529e5b" exitCode=0 Dec 03 14:30:02 crc kubenswrapper[4690]: I1203 14:30:02.154445 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" event={"ID":"41815399-ff92-43f3-af2d-3433fa3e73c9","Type":"ContainerDied","Data":"b9a6f484acadba9194e493665e9bbe4adb44ed90bfbdec8fc7c95feb38529e5b"} Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.589821 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.668364 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41815399-ff92-43f3-af2d-3433fa3e73c9-secret-volume\") pod \"41815399-ff92-43f3-af2d-3433fa3e73c9\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.668409 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41815399-ff92-43f3-af2d-3433fa3e73c9-config-volume\") pod \"41815399-ff92-43f3-af2d-3433fa3e73c9\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.668512 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq5sb\" (UniqueName: \"kubernetes.io/projected/41815399-ff92-43f3-af2d-3433fa3e73c9-kube-api-access-xq5sb\") pod \"41815399-ff92-43f3-af2d-3433fa3e73c9\" (UID: \"41815399-ff92-43f3-af2d-3433fa3e73c9\") " Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.669431 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41815399-ff92-43f3-af2d-3433fa3e73c9-config-volume" (OuterVolumeSpecName: "config-volume") pod "41815399-ff92-43f3-af2d-3433fa3e73c9" (UID: "41815399-ff92-43f3-af2d-3433fa3e73c9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.676717 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41815399-ff92-43f3-af2d-3433fa3e73c9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "41815399-ff92-43f3-af2d-3433fa3e73c9" (UID: "41815399-ff92-43f3-af2d-3433fa3e73c9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.676857 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41815399-ff92-43f3-af2d-3433fa3e73c9-kube-api-access-xq5sb" (OuterVolumeSpecName: "kube-api-access-xq5sb") pod "41815399-ff92-43f3-af2d-3433fa3e73c9" (UID: "41815399-ff92-43f3-af2d-3433fa3e73c9"). InnerVolumeSpecName "kube-api-access-xq5sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.771198 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq5sb\" (UniqueName: \"kubernetes.io/projected/41815399-ff92-43f3-af2d-3433fa3e73c9-kube-api-access-xq5sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.771239 4690 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41815399-ff92-43f3-af2d-3433fa3e73c9-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:03 crc kubenswrapper[4690]: I1203 14:30:03.771250 4690 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41815399-ff92-43f3-af2d-3433fa3e73c9-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:04 crc kubenswrapper[4690]: I1203 14:30:04.177693 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" event={"ID":"41815399-ff92-43f3-af2d-3433fa3e73c9","Type":"ContainerDied","Data":"26a0d31cb1c216376f3b542ddb03a6f75e48bf5202321318f3b067c2ef21c4f9"} Dec 03 14:30:04 crc kubenswrapper[4690]: I1203 14:30:04.178427 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26a0d31cb1c216376f3b542ddb03a6f75e48bf5202321318f3b067c2ef21c4f9" Dec 03 14:30:04 crc kubenswrapper[4690]: I1203 14:30:04.177744 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-lzvvk" Dec 03 14:30:04 crc kubenswrapper[4690]: I1203 14:30:04.667526 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr"] Dec 03 14:30:04 crc kubenswrapper[4690]: I1203 14:30:04.678782 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-n2wnr"] Dec 03 14:30:06 crc kubenswrapper[4690]: I1203 14:30:06.354554 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5717c7dc-0f59-490f-8b60-33b8275899b6" path="/var/lib/kubelet/pods/5717c7dc-0f59-490f-8b60-33b8275899b6/volumes" Dec 03 14:30:12 crc kubenswrapper[4690]: I1203 14:30:12.316764 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:30:12 crc kubenswrapper[4690]: E1203 14:30:12.317735 4690 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5fc8_openshift-machine-config-operator(e23e6be9-62d7-4b0a-a073-52e9fc3a8723)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" Dec 03 14:30:24 crc kubenswrapper[4690]: I1203 14:30:24.321731 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" Dec 03 14:30:25 crc kubenswrapper[4690]: I1203 14:30:25.422199 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"24f1a08bea0c44b5447a84ebc5e3ffa93d7f2d05cfd4d37ef79d3e8fcdc52591"} Dec 03 14:30:34 crc kubenswrapper[4690]: I1203 14:30:34.548591 4690 generic.go:334] "Generic (PLEG): container finished" podID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerID="887da7cca4071f6ea28acd15a8733ad0bbd8c38940c481168c8f804c5209a62b" exitCode=0 Dec 03 14:30:34 crc kubenswrapper[4690]: I1203 14:30:34.548676 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" event={"ID":"b36061d5-2fe4-4c05-8c07-02674c7337ec","Type":"ContainerDied","Data":"887da7cca4071f6ea28acd15a8733ad0bbd8c38940c481168c8f804c5209a62b"} Dec 03 14:30:34 crc kubenswrapper[4690]: I1203 14:30:34.549994 4690 scope.go:117] "RemoveContainer" containerID="887da7cca4071f6ea28acd15a8733ad0bbd8c38940c481168c8f804c5209a62b" Dec 03 14:30:35 crc kubenswrapper[4690]: I1203 14:30:35.370119 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksl6l_must-gather-vwxpt_b36061d5-2fe4-4c05-8c07-02674c7337ec/gather/0.log" Dec 03 14:30:43 crc kubenswrapper[4690]: I1203 14:30:43.624428 4690 scope.go:117] "RemoveContainer" containerID="abaf56525d7cdf3965da80f820ec2ca8381781de1172d5ee438d1f5d2ce161e0" Dec 03 14:30:43 crc kubenswrapper[4690]: I1203 14:30:43.652362 4690 scope.go:117] "RemoveContainer" containerID="169c8b15637f1ff2e65a6678b6f409d618ad651e29c1789f007e16c50208f91e" Dec 03 14:30:46 crc kubenswrapper[4690]: I1203 14:30:46.533968 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksl6l/must-gather-vwxpt"] Dec 03 14:30:46 crc kubenswrapper[4690]: I1203 14:30:46.534826 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" podUID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerName="copy" containerID="cri-o://90e73552df83a70985013aef2b413554dec9f4d024b05b689b42fcf4a4c5af54" gracePeriod=2 Dec 03 14:30:46 crc kubenswrapper[4690]: I1203 14:30:46.550659 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksl6l/must-gather-vwxpt"] Dec 03 14:30:46 crc kubenswrapper[4690]: I1203 14:30:46.680058 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksl6l_must-gather-vwxpt_b36061d5-2fe4-4c05-8c07-02674c7337ec/copy/0.log" Dec 03 14:30:46 crc kubenswrapper[4690]: I1203 14:30:46.680484 4690 generic.go:334] "Generic (PLEG): container finished" podID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerID="90e73552df83a70985013aef2b413554dec9f4d024b05b689b42fcf4a4c5af54" exitCode=143 Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.029910 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksl6l_must-gather-vwxpt_b36061d5-2fe4-4c05-8c07-02674c7337ec/copy/0.log" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.030959 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.098990 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b36061d5-2fe4-4c05-8c07-02674c7337ec-must-gather-output\") pod \"b36061d5-2fe4-4c05-8c07-02674c7337ec\" (UID: \"b36061d5-2fe4-4c05-8c07-02674c7337ec\") " Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.099081 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlqfm\" (UniqueName: \"kubernetes.io/projected/b36061d5-2fe4-4c05-8c07-02674c7337ec-kube-api-access-hlqfm\") pod \"b36061d5-2fe4-4c05-8c07-02674c7337ec\" (UID: \"b36061d5-2fe4-4c05-8c07-02674c7337ec\") " Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.107102 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36061d5-2fe4-4c05-8c07-02674c7337ec-kube-api-access-hlqfm" (OuterVolumeSpecName: "kube-api-access-hlqfm") pod "b36061d5-2fe4-4c05-8c07-02674c7337ec" (UID: "b36061d5-2fe4-4c05-8c07-02674c7337ec"). InnerVolumeSpecName "kube-api-access-hlqfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.201812 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlqfm\" (UniqueName: \"kubernetes.io/projected/b36061d5-2fe4-4c05-8c07-02674c7337ec-kube-api-access-hlqfm\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.275045 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b36061d5-2fe4-4c05-8c07-02674c7337ec-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b36061d5-2fe4-4c05-8c07-02674c7337ec" (UID: "b36061d5-2fe4-4c05-8c07-02674c7337ec"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.302627 4690 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b36061d5-2fe4-4c05-8c07-02674c7337ec-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.695538 4690 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksl6l_must-gather-vwxpt_b36061d5-2fe4-4c05-8c07-02674c7337ec/copy/0.log" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.696414 4690 scope.go:117] "RemoveContainer" containerID="90e73552df83a70985013aef2b413554dec9f4d024b05b689b42fcf4a4c5af54" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.696497 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksl6l/must-gather-vwxpt" Dec 03 14:30:47 crc kubenswrapper[4690]: I1203 14:30:47.720090 4690 scope.go:117] "RemoveContainer" containerID="887da7cca4071f6ea28acd15a8733ad0bbd8c38940c481168c8f804c5209a62b" Dec 03 14:30:48 crc kubenswrapper[4690]: I1203 14:30:48.327719 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b36061d5-2fe4-4c05-8c07-02674c7337ec" path="/var/lib/kubelet/pods/b36061d5-2fe4-4c05-8c07-02674c7337ec/volumes" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.788162 4690 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d9rq8"] Dec 03 14:30:59 crc kubenswrapper[4690]: E1203 14:30:59.789327 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerName="gather" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.789346 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerName="gather" Dec 03 14:30:59 crc kubenswrapper[4690]: E1203 14:30:59.789417 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41815399-ff92-43f3-af2d-3433fa3e73c9" containerName="collect-profiles" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.789426 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="41815399-ff92-43f3-af2d-3433fa3e73c9" containerName="collect-profiles" Dec 03 14:30:59 crc kubenswrapper[4690]: E1203 14:30:59.789446 4690 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerName="copy" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.789457 4690 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerName="copy" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.789688 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerName="gather" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.789701 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="41815399-ff92-43f3-af2d-3433fa3e73c9" containerName="collect-profiles" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.789728 4690 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36061d5-2fe4-4c05-8c07-02674c7337ec" containerName="copy" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.791686 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.807288 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9rq8"] Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.858450 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s29w2\" (UniqueName: \"kubernetes.io/projected/59625137-a691-43d4-8355-25e31fb213b0-kube-api-access-s29w2\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.858510 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-catalog-content\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.858607 4690 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-utilities\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.961228 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s29w2\" (UniqueName: \"kubernetes.io/projected/59625137-a691-43d4-8355-25e31fb213b0-kube-api-access-s29w2\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.961301 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-catalog-content\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.961384 4690 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-utilities\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.962578 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-utilities\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.962645 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-catalog-content\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:30:59 crc kubenswrapper[4690]: I1203 14:30:59.985839 4690 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s29w2\" (UniqueName: \"kubernetes.io/projected/59625137-a691-43d4-8355-25e31fb213b0-kube-api-access-s29w2\") pod \"redhat-marketplace-d9rq8\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:31:00 crc kubenswrapper[4690]: I1203 14:31:00.125721 4690 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:31:00 crc kubenswrapper[4690]: I1203 14:31:00.645365 4690 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9rq8"] Dec 03 14:31:00 crc kubenswrapper[4690]: I1203 14:31:00.970578 4690 generic.go:334] "Generic (PLEG): container finished" podID="59625137-a691-43d4-8355-25e31fb213b0" containerID="377006bdd46c89f0750ab5790b666f732166714dbfedbe09a73fd381d12f6239" exitCode=0 Dec 03 14:31:00 crc kubenswrapper[4690]: I1203 14:31:00.970644 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9rq8" event={"ID":"59625137-a691-43d4-8355-25e31fb213b0","Type":"ContainerDied","Data":"377006bdd46c89f0750ab5790b666f732166714dbfedbe09a73fd381d12f6239"} Dec 03 14:31:00 crc kubenswrapper[4690]: I1203 14:31:00.970938 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9rq8" event={"ID":"59625137-a691-43d4-8355-25e31fb213b0","Type":"ContainerStarted","Data":"577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115"} Dec 03 14:31:01 crc kubenswrapper[4690]: I1203 14:31:01.983968 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9rq8" event={"ID":"59625137-a691-43d4-8355-25e31fb213b0","Type":"ContainerStarted","Data":"70bdbbb1a768e8916a458f3a2036a7d35f285fd4652a36a7e65441215011bbb2"} Dec 03 14:31:03 crc kubenswrapper[4690]: I1203 14:31:03.001199 4690 generic.go:334] "Generic (PLEG): container finished" podID="59625137-a691-43d4-8355-25e31fb213b0" containerID="70bdbbb1a768e8916a458f3a2036a7d35f285fd4652a36a7e65441215011bbb2" exitCode=0 Dec 03 14:31:03 crc kubenswrapper[4690]: I1203 14:31:03.001273 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9rq8" event={"ID":"59625137-a691-43d4-8355-25e31fb213b0","Type":"ContainerDied","Data":"70bdbbb1a768e8916a458f3a2036a7d35f285fd4652a36a7e65441215011bbb2"} Dec 03 14:31:05 crc kubenswrapper[4690]: I1203 14:31:05.028276 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9rq8" event={"ID":"59625137-a691-43d4-8355-25e31fb213b0","Type":"ContainerStarted","Data":"57e0556eea1eb880f09d366389563a65589b53289b3e70995659c64ccd15cd7e"} Dec 03 14:31:05 crc kubenswrapper[4690]: I1203 14:31:05.061676 4690 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d9rq8" podStartSLOduration=2.809171192 podStartE2EDuration="6.06164791s" podCreationTimestamp="2025-12-03 14:30:59 +0000 UTC" firstStartedPulling="2025-12-03 14:31:00.97322926 +0000 UTC m=+4906.954149693" lastFinishedPulling="2025-12-03 14:31:04.225705978 +0000 UTC m=+4910.206626411" observedRunningTime="2025-12-03 14:31:05.050007224 +0000 UTC m=+4911.030927657" watchObservedRunningTime="2025-12-03 14:31:05.06164791 +0000 UTC m=+4911.042568343" Dec 03 14:31:10 crc kubenswrapper[4690]: I1203 14:31:10.126440 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:31:10 crc kubenswrapper[4690]: I1203 14:31:10.127290 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:31:10 crc kubenswrapper[4690]: I1203 14:31:10.186646 4690 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:31:11 crc kubenswrapper[4690]: I1203 14:31:11.146606 4690 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:31:11 crc kubenswrapper[4690]: I1203 14:31:11.209275 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9rq8"] Dec 03 14:31:13 crc kubenswrapper[4690]: I1203 14:31:13.112310 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d9rq8" podUID="59625137-a691-43d4-8355-25e31fb213b0" containerName="registry-server" containerID="cri-o://57e0556eea1eb880f09d366389563a65589b53289b3e70995659c64ccd15cd7e" gracePeriod=2 Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.127261 4690 generic.go:334] "Generic (PLEG): container finished" podID="59625137-a691-43d4-8355-25e31fb213b0" containerID="57e0556eea1eb880f09d366389563a65589b53289b3e70995659c64ccd15cd7e" exitCode=0 Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.127322 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9rq8" event={"ID":"59625137-a691-43d4-8355-25e31fb213b0","Type":"ContainerDied","Data":"57e0556eea1eb880f09d366389563a65589b53289b3e70995659c64ccd15cd7e"} Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.127782 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9rq8" event={"ID":"59625137-a691-43d4-8355-25e31fb213b0","Type":"ContainerDied","Data":"577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115"} Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.127801 4690 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115" Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.190914 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.230595 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s29w2\" (UniqueName: \"kubernetes.io/projected/59625137-a691-43d4-8355-25e31fb213b0-kube-api-access-s29w2\") pod \"59625137-a691-43d4-8355-25e31fb213b0\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.230714 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-utilities\") pod \"59625137-a691-43d4-8355-25e31fb213b0\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.230784 4690 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-catalog-content\") pod \"59625137-a691-43d4-8355-25e31fb213b0\" (UID: \"59625137-a691-43d4-8355-25e31fb213b0\") " Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.232117 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-utilities" (OuterVolumeSpecName: "utilities") pod "59625137-a691-43d4-8355-25e31fb213b0" (UID: "59625137-a691-43d4-8355-25e31fb213b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.262234 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59625137-a691-43d4-8355-25e31fb213b0-kube-api-access-s29w2" (OuterVolumeSpecName: "kube-api-access-s29w2") pod "59625137-a691-43d4-8355-25e31fb213b0" (UID: "59625137-a691-43d4-8355-25e31fb213b0"). InnerVolumeSpecName "kube-api-access-s29w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.265493 4690 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59625137-a691-43d4-8355-25e31fb213b0" (UID: "59625137-a691-43d4-8355-25e31fb213b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.351132 4690 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s29w2\" (UniqueName: \"kubernetes.io/projected/59625137-a691-43d4-8355-25e31fb213b0-kube-api-access-s29w2\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.351550 4690 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:14 crc kubenswrapper[4690]: I1203 14:31:14.351566 4690 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59625137-a691-43d4-8355-25e31fb213b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:15 crc kubenswrapper[4690]: I1203 14:31:15.136505 4690 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9rq8" Dec 03 14:31:15 crc kubenswrapper[4690]: I1203 14:31:15.163772 4690 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9rq8"] Dec 03 14:31:15 crc kubenswrapper[4690]: I1203 14:31:15.173951 4690 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9rq8"] Dec 03 14:31:16 crc kubenswrapper[4690]: I1203 14:31:16.329510 4690 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59625137-a691-43d4-8355-25e31fb213b0" path="/var/lib/kubelet/pods/59625137-a691-43d4-8355-25e31fb213b0/volumes" Dec 03 14:31:17 crc kubenswrapper[4690]: E1203 14:31:17.113393 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice/crio-577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:31:27 crc kubenswrapper[4690]: E1203 14:31:27.412006 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice/crio-577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115\": RecentStats: unable to find data in memory cache]" Dec 03 14:31:37 crc kubenswrapper[4690]: E1203 14:31:37.698181 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice/crio-577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:31:43 crc kubenswrapper[4690]: I1203 14:31:43.760943 4690 scope.go:117] "RemoveContainer" containerID="1b926664028e64d3be77ae616c1974ca9342c7752c896e3014b49134ec814ffd" Dec 03 14:31:47 crc kubenswrapper[4690]: E1203 14:31:47.964159 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice/crio-577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115\": RecentStats: unable to find data in memory cache]" Dec 03 14:31:58 crc kubenswrapper[4690]: E1203 14:31:58.233793 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice/crio-577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:32:08 crc kubenswrapper[4690]: E1203 14:32:08.528509 4690 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59625137_a691_43d4_8355_25e31fb213b0.slice/crio-577692c9487d2deae2398b1dec832760089ad7bd6817620fd2d15b60c0c28115\": RecentStats: unable to find data in memory cache]" Dec 03 14:32:46 crc kubenswrapper[4690]: I1203 14:32:46.823892 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:32:46 crc kubenswrapper[4690]: I1203 14:32:46.824598 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:33:16 crc kubenswrapper[4690]: I1203 14:33:16.824037 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:33:16 crc kubenswrapper[4690]: I1203 14:33:16.826855 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:33:46 crc kubenswrapper[4690]: I1203 14:33:46.823704 4690 patch_prober.go:28] interesting pod/machine-config-daemon-p5fc8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:33:46 crc kubenswrapper[4690]: I1203 14:33:46.825893 4690 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:33:46 crc kubenswrapper[4690]: I1203 14:33:46.825957 4690 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" Dec 03 14:33:46 crc kubenswrapper[4690]: I1203 14:33:46.827365 4690 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24f1a08bea0c44b5447a84ebc5e3ffa93d7f2d05cfd4d37ef79d3e8fcdc52591"} pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:33:46 crc kubenswrapper[4690]: I1203 14:33:46.827507 4690 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" podUID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerName="machine-config-daemon" containerID="cri-o://24f1a08bea0c44b5447a84ebc5e3ffa93d7f2d05cfd4d37ef79d3e8fcdc52591" gracePeriod=600 Dec 03 14:33:47 crc kubenswrapper[4690]: I1203 14:33:47.830319 4690 generic.go:334] "Generic (PLEG): container finished" podID="e23e6be9-62d7-4b0a-a073-52e9fc3a8723" containerID="24f1a08bea0c44b5447a84ebc5e3ffa93d7f2d05cfd4d37ef79d3e8fcdc52591" exitCode=0 Dec 03 14:33:47 crc kubenswrapper[4690]: I1203 14:33:47.830456 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerDied","Data":"24f1a08bea0c44b5447a84ebc5e3ffa93d7f2d05cfd4d37ef79d3e8fcdc52591"} Dec 03 14:33:47 crc kubenswrapper[4690]: I1203 14:33:47.830778 4690 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5fc8" event={"ID":"e23e6be9-62d7-4b0a-a073-52e9fc3a8723","Type":"ContainerStarted","Data":"19f33af1f6b6a2c878d151ac9fab7a92c2abb00533b1d6d7b0433104c7144fba"} Dec 03 14:33:47 crc kubenswrapper[4690]: I1203 14:33:47.830807 4690 scope.go:117] "RemoveContainer" containerID="9ec3cc1de8eb688a93e5bbaf3407e3d13ee90a28280b466f7ab8b6c2e88a8fbb" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114045220024437 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114045221017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114032657016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114032657015461 5ustar corecore